Next Article in Journal
Codazzi Tensors and the Quasi-Statistical Structure Associated with Affine Connections on Three-Dimensional Lorentzian Lie Groups
Previous Article in Journal
Automatic, Illumination-Invariant and Real-Time Green-Screen Keying Using Deeply Guided Linear Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Feature Combination-Based Graph Convolutional Neural Network Model for Relation Extraction

1
College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
2
School of Automation Science and Engineering, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Symmetry 2021, 13(8), 1458; https://doi.org/10.3390/sym13081458
Submission received: 15 June 2021 / Revised: 25 July 2021 / Accepted: 4 August 2021 / Published: 9 August 2021
(This article belongs to the Section Computer)

Abstract

:
The task to extract relations tries to identify relationships between two named entities in a sentence. Because a sentence usually contains several named entities, capturing structural information of a sentence is important to support this task. Currently, graph neural networks are widely implemented to support relation extraction, in which dependency trees are employed to generate adjacent matrices for encoding structural information of a sentence. Because parsing a sentence is error-prone, it influences the performance of a graph neural network. On the other hand, a sentence is structuralized by several named entities, which precisely segment a sentence into several parts. Different features can be combined by prior knowledge and experience, which are effective to initialize a symmetric adjacent matrix for a graph neural network. Based on this phenomenon, we proposed a feature combination-based graph convolutional neural network model (FC-GCN). It has the advantages of encoding structural information of a sentence, considering prior knowledge, and avoiding errors caused by parsing. In the experiments, the results show significant improvement, which outperform existing state-of-the-art performances.

1. Introduction

Relation extraction means extracting entity-relationship triples from unstructured text. For example, in “teams of nurses and doctors were seen in packed emergency rooms attending to the wounded”, the entity “packed emergency rooms” and the entity “the wounded” belong to a “physical (PHYS)” relationship. Relation extraction can transform unstructured text into structured data. It can promote the automatic construction of a knowledge base [1], understand user query intentions, and improve the search efficiency of search engines.
At present, neural network models are used by more and more researchers. It can well-capture the semantic information in a sentence. Because a sentence usually contains several named entities, it is also important to capture structural information of a sentence for supporting relation extraction. For example, He et al. [2] used latent structural information and semantic features to optimize the representation of discourse arguments, which can enhance the semantic understanding of implicit relationships. More recently, more and more researchers have been using graph neural network models for relation extraction, and have made a lot of progress. One of the important steps of using graph neural networks for relation extraction is to generate an adjacency matrix. Most researchers use dependency trees to generate adjacent matrices for encoding structural information of a sentence. For instance, Sun et al. [3] converted a sentence-based dependency tree into a directed graph to consider the structural information in the tree.
However, on the one hand, this kind of graph neural network model based on dependency trees has two problems. One is that it often has poor performance by inaccurate chunking or parsing; the other is that the prior knowledge that can be obtained through our previous experience is not convenient to be used. On the other hand, in our normal cognitive process, when we see a sentence, our brain will form a graph structure based on prior knowledge of the sentence. Then, after relevant processing by our brain, we can get the meaning of the sentence. More specifically, if this sentence contains two entities that we want to judge the relationship, we will get the relationship between these two entities in this sentence.
Moreover, a sentence is structuralized by two named entities, which precisely segment a sentence into several parts. Different features can be obtained by prior knowledge and experience, which are not only effective to initialize a graph for graph neural networks, but also capture the structural information of the sentence. Additionally, it is easy to cause the problem of feature sparseness for the relation extraction basically based on a sentence with a few words. In order to solve the problem, Chen et al. [4] constructed a set space model, which uses language characteristics to combine the features in the sentence into different sets. Thus, in this paper, we also use this method to obtain more features of the sentence. Moreover, we make some improvements on the basis of the atomic features and combined features in Chen et al. [4]. Next, we use the combined features to initialize a graph and apply it to the graph convolutional neural network (GCN). Therefore, a feature combination-based graph convolutional neural network model for relation extraction is proposed by us. In this way, we not only make good use of sentence structural information and prior knowledge, but also avoid the above-mentioned problems caused by the dependency trees.
We tested the performance of the model on the ACE05 English dataset, CoNLL04 dataset, and SciERC dataset. Experiments show that our model can achieve better performance for these datasets.
The contributions of this paper can be divided into the following points:
  • Our methods to generate combined features are used to initialize an adjacent matrix for a graph neural network. It is effective at capturing the structural information of a sentence.
  • Based on a graph convolutional neural network, a deep architecture is designed to support relation extraction. It outperforms existing state-of-the-art performance.
The rest of this paper is organized as follows. Section 2 shows related work. Section 3 elaborates on the construction process of our model. The details related to the experiment are shown in Section 4. Section 5 presents our conclusion and future work.

2. Related Work

At present, the neural network model has achieved great improvement in relation extraction tasks. The main neural networks used are the Convolutional Neural Network (CNN) [5,6,7] and Recurrent Neural Network (RNN) [8,9,10]. Moreover, according to whether the extracted entity pairs span sentences, the relation extraction model can be divided into a sentence-level relation extraction model [11] and cross-sentence relation extraction model [12]. There are also some researchers that perform document-level relation extraction [13,14]. Because an entity pair may have multiple relationships, some people conduct research on an entity pair corresponding to only one relationship, while some people take into account the existence of multiple relationships in the entity pair for relation extraction [15]. In this paper, we only consider one entity pair corresponding to one relationship.
In the deep learning framework, because the neural network model is weak in obtaining the structural information of the sentence, position information is often used to capture sentence structural information. For example, Zeng et al. [7] added position information for distant supervision relation extraction. Veyseh et al. [10] not only considers the word embedding and position embedding, but also adds the entity tag embedding for relation extraction. Shen and Huang [16] combines the attention mechanism and position information to obtain sentence structural information.
More recently, many researchers have used sentence-based dependency trees to generate a graph for encoding structural information of the sentence, then to apply a graph neural network to perform relation extraction. One of the important steps to construct a graph neural network model is to construct an adjacency matrix, or in other words, to obtain a graph. At present, Guo et al. [17], Vashishth et al. [18] and Fu and Ma [19] inputted the entire tree into graph convolutional neural networks to perform relation extraction. Zhang et al. [20] used two entities in the Least Common Ancestor (LCA) subtree of the dependency tree to construct the graph with nodes that do not exceed the K distance. In view of the above construction of a complete graph based on the dependency tree, the structural information in the dependency tree is not considered. Sun et al. [3] established a directed weighted graph while retaining the structure of the information depending on the tree.
In addition to using dependency trees to construct graphs of sentences, there are others who directly construct graphs on the sentences. Vashishth et al. [21] regard entities as nodes, and construct a multi-relation graph to predict multiple relations between entities. Sun et al. [22] use entities and entity pairs to construct a bipartite graph for joint extraction of entities and relationship types. However, the above-mentioned processes of constructing a graph cannot make good use of prior knowledge. Moreover, when the graph is obtained using a dependency tree, because the dependency tree is generated by an external tool (e.g., StanfordCoreNLP) which is error-prone, this may lead to poor performance of the model. Therefore, in our experiments, we do not use dependency trees to build graphs, but to build undirected graphs based on features related to two entities, which avoid the problems caused by dependency trees.

3. Model

In this part, we will introduce the FC-GCN model in detail. The overall model diagram is shown in Figure 1.

3.1. The Atomic and Combined Features

Combined features can well-capture the structural information in sentences [23] and are widely used in relation extraction tasks. It is generally constructed by using syntactic and semantic rules. In this paper, we have made further improvements on the basis of the features related to two entities in Chen et al. [4]. Below, we first introduce the features which we used in Chen et al. [4], and it includes the entity type, entity subtype, part of speech (POS) of the word on the left and right of the entity, entity heads tag, and relative positions of two entities. Additionally, we also regard these features as atomic features.
Given a sentence s = [ w 1 , w 2 , . . . ] ( w j is a word) and two entities e 1 and e 2 in it, let e i represent the i-th entity ( e 1 or e 2 ). In order to obtain the above-mentioned atomic features, we assume that the functions RightPosOf( e i ), LeftPosOf( e i ), TypeOf( e i ), SubTypeOf( e i ), HeadOf( e i ), PositionBetween( e 1 , e 2 ), respectively, can obtain the POS tags of the word on the right and left of e i , the type of e i , the subtype of e i , the head tag of e i , and the relative position of e 1 and e 2 . In that way, we can get 11 atomic features for every two entities as follows:
x 1 = R i g h t P o s O f ( e 1 ) x 7 = S u b T y p e O f ( e 1 ) x 2 = L e f t P o s O f ( e 1 ) x 8 = S u b T y p e O f ( e 2 ) x 3 = R i g h t P o s O f ( e 2 ) x 9 = H e a d O f ( e 1 ) x 4 = L e f t P o s O f ( e 2 ) x 10 = H e a d O f ( e 2 ) x 5 = T y p e O f ( e 1 ) x 11 = P o s i t i o n B e t w e e n ( e 1 , e 2 ) x 6 = T y p e O f ( e 2 )
where x k represents the k-th atomic feature. In addition to POS tags and relative position features, the remaining other features are marked in the corresponding corpus. As for the POS tags and relative position features, we obtained them through external tools (e.g., nltk) and manual annotations, respectively. Moreover, there are four kinds of relative position information, which are e 1 in front of e 2 (marked as 1), e 2 in front of e 1 (marked as 2), e 1 contains e 2 (marked as 3), and e 2 contains e 1 (marked as 4).
Because we consider that two entities are also crucial for judging the relationship between entities, we also regard the two entities as atomic features. Moreover, we assume that the functions Entity( e 1 ) and Entity( e 1 ), respectively, can obtain e 1 and e 2 . Then we can get the following two new atomic features.
x 12 = E n t i t y ( e 1 ) x 13 = E n t i t y ( e 2 )
According to Chen et al. [4], based on the above atomic features and prior knowledge, the eight combined features are designed as follows:
X 1 = R i g h t P o s O f ( e 1 ) T y p l e O f ( e 1 ) X 2 = L e f t P o s O f ( e 1 ) T y p l e O f ( e 1 ) X 3 = R i g h t P o s O f ( e 2 ) T y p l e O f ( e 2 ) X 4 = L e f t P o s O f ( e 2 ) T y p l e O f ( e 2 ) X 5 = T y p l e O f ( e 1 ) T y p l e O f ( e 2 ) X 6 = S u b T y p e O f ( e 1 ) S u b T y p e O f ( e 2 ) X 7 = H e a d O f ( e 1 ) H e a d O f ( e 2 ) X 8 = P o s i t i o n B e t w e e n ( e 1 , e 2 )
where ⊕ is the connection symbol, and X t represents the t-th combined feature. According to each combination feature, we can get the corresponding combination rule.
In this paper, we directly regard x 12 and x 13 as combined features. It is equivalent to using these two atomic features to construct a node of the graph without constructing an edge among it with any other node. Then the two new combined features are as follows:
X 9 = E n t i t y ( e 1 ) X 10 = E n t i t y ( e 2 )
In summary, we can get 13 atomic features ( x 1 , . . . , x 13 ) and 10 combined features ( X 1 , . . . , X 10 ) for every two entities. So far, we can see that x 11 , x 12 and x 13 are both atomic and combined features. In this paper, these atomic features and combined features are our prior knowledge. Furthermore, rules or patterns to generate combined features are also regarded as “prior knowledge”. However, if there are not so many atomic features in the dataset, this number may be reduced, but the construction process of the entire model remains unchanged.

3.2. Graph Based on Combined Features

After we get the atomic features and combined features related to the two entities, the graph based on the combined features can be constructed. Specifically, we use these atomic features as nodes and construct edges between nodes according to the combination rules ( X 1 , . . . , X 10 ) to obtain the graphs based on combined features. Then we use graph convolutional neural networks to extract higher-dimensional feature representations of this graph. For example, in this way, a total of 13 atomic features ( x 1 , . . . , x 13 ) can be generated on the ACE05 English dataset. Then we can construct an undirected graph with 13 vertices. The sample diagram is shown in Figure 2.
Formally, we define V as the set of nodes and E as the set of edges. Whether an edge is constructed between two nodes depends on the combination rules. In other words, if two nodes belong to one of the combined features, then they are connected with undirected edges. Otherwise, the edges are not constructed. Then, we get the undirected graph G = ( V , E ) . Moreover, we assume that A is the adjacency matrix of the graph. If there exist an edge from node i to node j, then we define A i j = A j i = 1 , otherwise, A i j = A j i = 0 . We add a self-loop for each node in the graph, so the values on the diagonal of A are all 1.

3.3. GCN Module

Inspired by the convolutional neural network, Kipf and Welling [24] built a graph convolutional neural network on the graph, which can update the information of the current node according to the information of the neighbor nodes. After multiple layers of GCN, we can think that each node has collected information about other nodes in the entire graph. Specifically, given the undirected graph G = ( V , E ) based on combined features with n nodes, in an L-layer GCN, each layer can be expressed as a nonlinear equation:
H ( l + 1 ) = σ ( D ˜ 1 2 A ˜ D ˜ 1 2 H ( l ) W ( l ) ) ,
where H ( l + 1 ) is the node representation of the l-th layer, A ˜ = A + I . I is the identity matrix. D ˜ i i = j A ˜ i j is the degree matrix. W is the parameter matrix participating in training. σ is the activation function (e.g., relu). Until now, we have obtained a graph based on combined features and a symmetric adjacency matrix A . Next, we embed the vertices ( x 1 , . . . , x 13 ) in the graph and denote it with X .
X = Embedding ( [ x 1 , . . . x 13 ] )
Then, H ( 0 ) = X . H ( 0 ) is the vector representation of the initial input to GCN. The output H ( L ) of the last layer is the updated feature matrix of each node.
In order to obtain a higher-dimensional feature representation, we can use multi-layer GCN:
H ( l + 1 ) = GCN ( H ( l ) )
In this way, each vertex can obtain information about other vertices in the entire graph. Finally, we get the feature representation h g c n of our graph based on combined features.
h g c n = H ( L )

3.4. Prediction

After applying the GCN model to the graph based on combined features, we get the feature representation ( h g c n ) of the vertex. Given the feature representation, our purpose is to determine what kind of relationship the two entities belong to in this sentence. Therefore, we apply a feedforward neural network (FFNN) and softmax layer on h g c n to make the prediction:
y ˜ = S o f t M a x ( F F N N ( h g c n ) ) ,
where y ˜ refers to the predicted category probability distribution. The loss function L is designed to minimize the cross entropy value:
L ( Θ ) = y l o g ( y ˜ ) ( 1 y ) l o g ( 1 y ˜ )
where y is the true category probability distribution, and Θ is the parameter to be learned in the entire neural network.

4. Experiment

4.1. Datasets

In the experiment, we used three datasets: ACE05, CoNLL04, and SciERC. All datasets are divided into the training set, validation set, and test set, according to the ratio of 8:1:1. The data analysis of each dataset is summarized in the Table 1. Below, we give a detailed introduction of each dataset:
ACE05: The ACE2005 English dataset is a standard dataset used for relation extraction, which is obtained from broadcasting, news, and web logs. It contains a total of seven entity types and six relationship types, and each entity has subtypes and head tags in addition to the type. The English dataset contains 506 documents and 6583 positive data. After adding negative data, we get 110,852 data.
CoNLL04: The data in the CoNLL04 dataset [25] are obtained from news articles. Each sentence contains entities and their corresponding relationships, and each entity is also marked with a type. There are four types of entities and five types of relationships in total. This dataset contains a total of 2048 positive data. After we add negative examples, there are a total of 19,080 data.
SciERC: The SciERC dataset [26] comes from abstracts of 500 AI papers. Like the CoNLL04 dataset, each sentence in it is marked with entities and their corresponding relationships, and each entity is also marked with types. There are six entity types and seven relationship types. There are 4648 positive cases, and 24,426 positive and negative cases.

4.2. Experimental Setting

In the experiment of the ACE05 dataset, because it contains the subtype and the head tag of the entity, there are 13 nodes in the final graph based on this dataset. In the CoNLL04 and SciERC datasets, because they do not contain these two features, the graph based on these two datasets has only nine nodes. Whether it is a thirteen-node graph or a nine-node graph, they all construct edges according to the combination rules described above to construct the graph.
In the experiment, we adjust the hyperparameters based on the validation set. Finally, the maximum sentence length is set to 100, the number of convolution kernels in CNN is 60, and the size of the convolution kernel is set to 7, the dropout ratio is set to 0.5 , the L2 regularization lambda is set to 1e-5, the initial learning rate is 1, and the decay rate reduced by 0.9 can get the best results. We also use the standard precision rate (P), recall rate (R) and F1 value as the evaluation indicators of the model. For each dataset, we have published its macro average F1 value.

4.3. Results on the Three Datasets

In this section, we show the multi-classification results of our model on ACE05, CoNLL04, and SciERC datasets in Table 2 and Table 3. In order to obtain information about the sentence itself containing two entities, we also add sentence features to our model. The sentence features are obtained through a convolutional neural network. From the Table 2 and Table 3, on the one hand, there are some improvements in the three data sets after adding sentence features. On the other hand, it can be seen from these two tables that the F1 value of the category “PHYS”, “Feature-of” are the lowest among all categories on the ACE05 and SciERC datasets, respectively. Additionally, the F1 value of the category “OrgBased_In” is the highest among all categories in the CoNLL04 dataset. Thus, for these features we used, they can well-capture the information of the entity pairs belonging to the category “OrgBased_In”, so as to well-separate the entity pairs belonging to this category on the CoNLL04 dataset. However, for the entity pair belonging to the category “PHYS” and “Feature-of”, it cannot extract the information well, resulting in the F1 value in this category being much lower than other categories on the ACE05 and SciERC datasets, respectively. Therefore, in future research, we need to pay more attention to the entity pairs in these categories, observe the characteristics of them, and see if there are other methods to extract the information of the entity pairs in this category.
In addition, by observing the results on the three datasets, it can be seen that the results on the ACE05 and CoNLL04 datasets are higher than the results on the SciERC dataset, and the results of the ACE05 and CoNLL04 datasets are not much different. This may be caused by the different sources of the three datasets. Through Section 4.1, we can find that the data in the ACE05 and CoNLL04 datasets are news-related, while the data in the SciERC dataset are abstracts from AI papers.

4.4. Comparison with Benchmark Models

We make a comparison with two benchmark models: An atomic feature model based on a neural network ( M a t o ) and a combined feature model based on a neural network ( M c o m ).
M a t o directly uses the atomic features mentioned above when a neural network performs relation extraction. It differs from the FC-GCN model by only one GCN module. First, we get the atomic features like the FC-GCN model. Then, we use the classifier to predict the relationship category. Since the difference between the M a t o and FC-GCN models is whether to use the GCN module, by comparing with this model, the importance of the GCN module to the FC-GCN model can be reflected.
M c o m uses the neural network to perform relation extraction after combining atomic features according to the combination rules. The difference between the M c o m model and the M a t o model is whether to combine the atomic features according to the above-mentioned combination rules to obtain the combined features. First, like the M a t o model, we get the atomic features. Then we use combination rules to transform atomic features into combination features. Finally, we use the classifier to predict the relationship category. Compared with the FC-GCN model, it lacks the GCN module, but has an additional feature combination step. Because the GCN module in the FC-GCN model has a certain function of combining features, comparing the M c o m model with the FC-GCN model can reflect that the GCN module has stronger capabilities, not just the ability to combine features.
Table 4 shows the comparison results of our model, the M a t o m i c model, and M c o m b i n e d model. It can be seen from the Table 4 that our model exceeds the M a t o m i c model and the M c o m b i n e d model in the three datasets. It indicates that the GCN module cannot only extract features in the graph effectively, but also extract more hidden information than simply combine features. More importantly, our model not only has good superiority in the F1 value, but also reach new heights in the accuracy and recall rate. In order to obtain information about the sentence itself containing two entities, we also add sentence features to our model. The sentence features are obtained by implementing a convolutional neural network on a sentence containing two entities. Then, the output is concatenated with the final feature representation and h gcn for classification. The result is shown in Table 4. After adding sentence features to our model, the results are not only improved, but the distribution of results on all models is the same as without sentence features.

4.5. Analysis

In summary, no matter how the adjacency matrix changes, our model is always better than the M c o m model and the M c o m model. It shows that the method of constructing the feature into a graph and then using the graph convolutional neural network to extract the features of the graph is effective.

4.6. Comparison with Related Works

In this part of the experiment, we compared the FC-GCN model with other strategies. Additionally, we show the results of the model (BERT_FC-GCN+sen) after BERT [27] pre-training. In the “BERT_FC-GCN+sen” model, the node’s vector representation X and the sentence-embedding containing two entities are both the outputs passing through the BERT layer.
Table 5 shows the comparison results between our model and other models on the ACE05 dataset. DRPC indirectly uses the dependency tree to predict the dependency relationship between words and the relationship between entities, so as to facilitate the capture of text information outside of syntactic information and the generalization of cross-domains. GCN(D+H) regards entities and entity pairs as nodes and constructs a bipartite graph. BERT-Z&H adds tags on both sides of the entity for relation extraction. Through this table, we can find that our model is better than these models. More specifically, our model is at least 1.91% higher than the above model in F1 value. It shows that the use of the graphs based on combined features can better extract the information between atomic features, so as to better extract the information between entity pairs, and finally make better decisions.
For the CoNLL04 and SciERC datasets, we show the results of comparison with other methods in Table 6. It can be seen from this table that our model is also higher than other methods. Additionally, the F1 value is at least 4.83% and 1.06% higher on the CoNLL04 and SciERC datasets, respectively.
Because relation extraction is implemented at sentence level, where a sentence usually contains a limited number of words, it suffers from a serious feature sparsity problem. The BERT embedding is pre-trained on external resources with the unsupervised method. It is effective to learn word semantic representations. They are helpful to reduce the feature sparsi-ty problem. Therefore, on these three data sets, the best results have been achieved after adding the BERT embedding.
In this section, we analyze the effect of different layers of GCN on the FC-GCN model. Specifically, we test the GCN layer numbers of 1, 2, and 3 on the three datasets. Additionally, the experimental results are shown in Figure 3. Firstly, it can be seen from the figure that no matter which dataset is used, the result of one layer of GCN is always higher than that of multi-layer GCN. Specifically, it can be clearly seen from the figure that the gap between one layer of GCN and two layers of GCN is the largest on the SciERC dataset, with a difference of 3.82% in F1 value. However, on the ACE05 and CoNLL04 datasets, the largest difference in F1 values is the two layers of GCN and the three layers of GCN, with a difference of 2.81%, 14.74% in F1 value, respectively. Moreover, compared with the results of the ACE05 and CoNLL04 datasets, the difference between the results on the SciERC dataset is not very large. This may be because the sources of these three datasets are different. Additionally, the information contained in them is also different. Thus, the hidden features needed to analyze the information are also different.

5. Conclusions and Future Work

In this paper, we proposed a novel feature combination-based graph convolutional neural network model (FC-GCN) for relation extraction, which no longer uses dependency trees to build graphs and reduces problems caused by external tools. Instead, we created an undirected graph based on combined features, which used atomic features as nodes and constructed edges between nodes according to the combination rules. Then, we applied a graph convolutional neural network to this graph and extracted high-dimensional information of this graph. Our model not only makes good use of sentence structural information and prior knowledge, but also avoids the problems caused by the dependency trees. Additionally, we demonstrated its superiority on three datasets. It outperforms existing state-of-the-art performances. Since we constructed a static graph in this paper, the adjacency matrix is fixed, and we will try to construct a dynamic graph in future research and test whether it will improve the experimental results. In addition, we will explore more ways to construct graphs in future research to better capture the structural information of sentences.

Author Contributions

J.X.: data curation, investigation, methodology, resources and writing (original draft); Y.C.: conceptualization, formal analysis, methodology, supervision and writing (review) and editing; Y.Q.: formal analysis, funding acquisition, project administration and supervision; R.H.: funding acquisition and supervision; Q.Z.: supervision and writing (review) and editing. All authors have read and agreed to the final version of the manuscript.

Funding

This work is supported by the Joint Funds of the National Natural Science Foundation of China under Grant No. U1836205, the Major Research Program of National Natural Science Foundation of China under Grant No. 91746116, National Natural Science Foundation of China under Grant No. 62066007 and No. 62066008, the Major Special Science and Technology Projects of Guizhou Province under Grant No. [2017]3002, the Key Projects of Science and Technology of Guizhou Province under Grant No. [2020] 1Z055 and Project of Guizhou Province Graduate Research Fund (Qianjiaohe YJSCXJH[2019]102).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Y.; Zhong, V.; Chen, D.; Angeli, G.; Manning, C.D. Position-aware attention and supervised data improve slot filling. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017. [Google Scholar]
  2. He, R.; Wang, J.; Guo, F.; Han, Y. Transs-driven joint learning architecture for implicit discourse relation recognition. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  3. Sun, K.; Zhang, R.; Mao, Y.; Mensah, S.; Liu, X. Relation extraction with convolutional network over learnable syntax-transport graph. Proc. AAAI Conf. Artif. 2020, 34, 8928–8935. [Google Scholar] [CrossRef]
  4. Chen, Y.; Wang, G.; Zheng, Q.; Qin, Y.; Chen, P. A set space model to capture structural information of a sentence. IEEE Access 2019, 7, 142515–142530. [Google Scholar] [CrossRef]
  5. Zeng, D.; Liu, K.; Lai, S.; Zhou, G.; Zhao, J. Relation classification via convolutional deep neural network. In Proceedings of the COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers; Dublin City University and Association for Computational Linguistics: Dublin, Ireland, 2014; pp. 2335–2344. Available online: https://www.aclweb.org/anthology/C14-1220 (accessed on 1 April 2021).
  6. Xu, K.; Feng, Y.; Huang, S.; Zhao, D. Semantic relation classification via convolutional neural networks with simple negative sampling. Comput. Sci. 2015, 71, 941–949. [Google Scholar]
  7. Zeng, D.; Liu, K.; Chen, Y.; Zhao, J. Distant supervision for relation extraction via piecewise convolutional neural networks. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015. [Google Scholar]
  8. Yan, X.; Mou, L.; Li, G.; Chen, Y.; Jin, Z. Classifying relations via long short term memory networks along shortest dependency paths. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP), Lisbon, Portugal, 17–21 September 2015. [Google Scholar]
  9. Zhang, Z.; Shu, X.; Yu, B.; Liu, T.; Guo, L. Distilling knowledge from well-informed soft labels for neural relation extraction. Proc. AAAI Conf. Artif. 2020, 34, 9620–9627. [Google Scholar]
  10. Veyseh, A.P.B.; Dernoncourt, F.; Dou, D.; Nguyen, T.H. Exploiting the syntax-model consistency for neural relation extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  11. Alt, C.; Gabryszak, A.; Hennig, L. Probing linguistic features of sentence-level representations in neural relation extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 1534–1545. Available online: https://www.aclweb.org/anthology/2020.acl-main.140 (accessed on 1 April 2021).
  12. Yu, D.; Sun, K.; Cardie, C.; Yu, D. Dialogue-based relation extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  13. Zhou, W.; Huang, K.; Ma, T.; Huang, J. Document-level relation extraction with adaptive thresholding and localized context pooling. arXiv 2020, arXiv:2010.11304. [Google Scholar]
  14. Jain, S.; van Zuylen, M.; Hajishirzi, H.; Beltagy, I. Scirex: A challenge dataset for document-level information extraction. arXiv 2020, arXiv:2005.00512. [Google Scholar]
  15. Wei, Z.; Su, J.; Wang, Y.; Tian, Y.; Chang, Y. A novel cascade binary tagging framework for relational triple extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  16. Shen, Y.; Huang, X.-J. Attention-based convolutional neural network for semantic relation extraction. In Proceedings of the COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, Japan, 11–16 December 2016; pp. 2526–2536. [Google Scholar]
  17. Guo, Z.; Zhang, Y.; Lu, W. Attention guided graph convolutional networks for relation extraction. arXiv 2019, arXiv:1906.07510. [Google Scholar]
  18. Vashishth, S.; Joshi, R.; Prayaga, S.S.; Bhattacharyya, C.; Talukdar, P. Reside: Improving distantly-supervised neural relation extraction using side information. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018. [Google Scholar]
  19. Fu, T.J.; Ma, W.Y. Graphrel: Modeling text as relational graphs for joint entity and relation extraction. ACL 2019, 1409–1418. [Google Scholar] [CrossRef] [Green Version]
  20. Zhang, Y.; Qi, P.; Manning, C.D. Graph convolution over pruned dependency trees improves relation extraction. arXiv 2018, arXiv:1809.10185. [Google Scholar]
  21. Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P.P. Composition-based multi-relational graph convolutional networks. arXiv 2019, arXiv:1911.03082. [Google Scholar]
  22. Sun, C.; Gong, Y.; Wu, Y.; Gong, M.; Duan, N. Joint type inference on entities and relations via graph convolutional networks. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar]
  23. Chen, Y.; Zheng, Q.; Zhang, W. Omni-word feature and soft constraint for chinese relation extraction. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Baltimore, MD, USA, 22–27 June 2014. [Google Scholar]
  24. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  25. Roth, D.; Yih, W.-T. A linear programming formulation for global inference in natural language tasks. In Proceedings of the Eighth Conference on Computational Natural Language Learning (CoNLL-2004) at HLT-NAACL 2004, Boston, MA, USA, 6–7 May 2004; Association for Computational Linguistics: Boston, MA, USA, 2004; pp. 1–8. Available online: https://www.aclweb.org/anthology/W04-2401 (accessed on 1 April 2021).
  26. Luan, Y.; He, L.; Ostendorf, M.; Hajishirzi, H. Multi-task identification of entities, relations, and coreferencefor scientific knowledge graph construction. arXiv 2018, arXiv:1808.09602. [Google Scholar]
  27. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  28. Kambhatla, N. Combining lexical, syntactic, and semantic features with maximum entropy models for extracting relations. In Proceedings of the ACL 2004 on Interactive Poster and Demonstration Sessions, ACLdemo ’04, Barcelona, Spain, 21–26 July 2004; Association for Computational Linguistics: Boston, MA, USA, 2004. [Google Scholar] [CrossRef]
  29. Zhou, G.; Su, J.; Zhang, J.; Zhang, M. Exploring various knowledge in relation extraction. In Proceedings of the ACL 2005, 43rd Annual Meeting of the Association for Computational Linguistics, Ann Arbor, MI, USA, 25–30 June 2005. [Google Scholar]
  30. Gormley, M.R.; Yu, M.; Dredze, M. Improved relation extraction with feature-rich compositional embedding models. arXiv 2015, arXiv:1505.02419. [Google Scholar]
  31. Veyseh, A.P.B.; Nguyen, T.H.; Dou, D. Improving cross-domain performance for relation extraction via dependency prediction and information flow control. arXiv 2019, arXiv:1907.03230. [Google Scholar]
  32. Wang, H.; Tan, M.; Yu, M.; Chang, S.; Wang, D.; Xu, K.; Guo, X.; Potdar, S. Extracting multiple-relations in one-pass with pre-trained transformers. arXiv 2019, arXiv:1902.01030. [Google Scholar]
  33. Zhong, Z.; Chen, D. A frustratingly easy approach for joint entity and relation extraction. arXiv 2020, arXiv:2010.12812. [Google Scholar]
  34. Chen, Y.; Wang, K.; Yang, W.; Qing, Y.; Huang, R.; Chen, P. A multi-channel deep neural network for relation extraction. IEEE Access 2020, 8, 13195–13203. [Google Scholar] [CrossRef]
  35. Chen, Y.; Yang, W.; Wang, K.; Qin, Y.; Huang, R.; Zheng, Q. A neuralized feature engineering method for entity relation extraction. Neural Netw. 2021, 141, 249–260. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overview of our FC-GCN model.
Figure 1. Overview of our FC-GCN model.
Symmetry 13 01458 g001
Figure 2. (a) The graph a constructed from the sentence “, which operates in 34 countries”. The two entities contained in this sentence are “which” ( e 1 ) and “34 countries” ( e 2 ), where X t and x r represents the t-th combination feature and the r-th atomic feature, respectively. From the positional relationship of these two entities in the sentence, it can be seen that e 1 is before e 2 , so its relative position relationship ( x 11 ) is marked as 1. (b) The values on the horizontal axis of the adjacency matrix b correspond to the atomic features on the vertical axis.
Figure 2. (a) The graph a constructed from the sentence “, which operates in 34 countries”. The two entities contained in this sentence are “which” ( e 1 ) and “34 countries” ( e 2 ), where X t and x r represents the t-th combination feature and the r-th atomic feature, respectively. From the positional relationship of these two entities in the sentence, it can be seen that e 1 is before e 2 , so its relative position relationship ( x 11 ) is marked as 1. (b) The values on the horizontal axis of the adjacency matrix b correspond to the atomic features on the vertical axis.
Symmetry 13 01458 g002
Figure 3. The results of different gcn layers on each dataset.
Figure 3. The results of different gcn layers on each dataset.
Symmetry 13 01458 g003
Table 1. Details of datasets used. Please see Section 4.1 for more details.
Table 1. Details of datasets used. Please see Section 4.1 for more details.
SplitACE05CoNLL04SciERC
train832931526419540
dev1377919082443
test1378019082443
Table 2. Multi-classification results of ACE05 dataset.
Table 2. Multi-classification results of ACE05 dataset.
DatasetRelationship TypeFC-GCNFC-GCN+sen
P(%)R(%)F1(%)P(%)R(%)F1(%)
ACE05PHYS68.2244.7954.0771.7046.6356.51
ART87.1853.1266.0293.5545.3161.05
GEN-AFF69.7744.7854.5561.2956.7258.91
ORG-AFF85.6377.3081.2591.2272.9781.08
PART-WHOLE73.6377.0175.2875.5874.7175.14
PER-SOC88.8970.0078.3291.6768.7578.57
total78.8961.1768.9180.8360.8569.43
Table 3. Multi-classification results of on the CoNLL04 and SciERC datasets.
Table 3. Multi-classification results of on the CoNLL04 and SciERC datasets.
DatasetRelationship TypeFC-GCNFC-GCN+sen
P(%)R(%)F1(%)P(%)R(%)F1(%)
CoNLL04Work_For62.7987.1072.9765.0083.8773.24
Kill70.5980.0075.0071.8876.6774.19
OrgBased_In75.5177.0876.2982.6179.1780.85
Live_In63.9369.6466.6759.0969.6463.93
Located_In65.8562.7964.2955.5681.4066.04
total67.7475.3271.3366.8378.1572.05
SciERCUsed-for35.0527.3130.7027.6735.3431.04
Feature-of00.0000.0000.0008.0010.5309.09
Hyponym-of76.1933.3346.3866.6737.5048.00
Evaluate-for55.1741.0347.0644.6853.8548.84
Part-of16.6704.5507.1419.4431.8224.14
Compare44.4420.0027.5920.0020.0020.00
Conjunction52.8347.4650.0032.3537.2934.65
total40.0524.8130.6431.2632.3331.79
Table 4. Comparison results with the M a t o model and M c o m model on three datasets. The meaning of “+sen” is to add a sentence feature containing two entities.
Table 4. Comparison results with the M a t o model and M c o m model on three datasets. The meaning of “+sen” is to add a sentence feature containing two entities.
ModelACE05CoNLL04SciERC
P(%)R(%)F1(%)P(%)R(%)F1(%)P(%)R(%)F1(%)
M a t o 77.0150.4860.9871.3065.0268.0252.1819.2628.14
M c o m 72.1154.4562.0572.2865.6368.7961.1519.7129.81
FC-GCN78.8961.1768.9167.7475.3271.3340.0524.8130.64
M a t o +sen64.6556.6260.3769.0467.1268.0733.4122.1326.62
M c o m b +sen74.7251.3660.8865.6672.9569.1130.2024.3326.95
FC-GCN+sen80.8360.8569.4366.8378.1572.0531.2632.3331.79
Table 5. Comparison results with other models on the ACE05 dataset.
Table 5. Comparison results with other models on the ACE05 dataset.
ModelP(%)R(%)F1(%)
K [28]63.5045.2052.80
G&J [29]77.2060.7068.00
FCM [30]71.5249.3258.26
DRPC [31]72.1063.4967.52
GCN(D+H) [22]68.765.467.00
BERT-EA [32]--67.46
BERT-Z&H [33]--73.10
FC-GCN78.8961.1768.91
FC-GCN+sen80.8360.8569.43
BERT_FC-GCN+sen85.6475.9580.50
Table 6. Comparison results with other models on the CoNLL04 and SciERC datasets.
Table 6. Comparison results with other models on the CoNLL04 and SciERC datasets.
ModelCoNLL04SciERC
P(%)R(%)F1(%)P(%)R(%)F1(%)
Multi-channel [34]74.5358.3864.9422.7818.3720.29
F a t o m i c [35]46.8973.4357.2320.3118.3019.25
F c o m b i n e d [35]72.8762.3967.2236.1626.7230.73
FC-GCN+sen66.8378.1572.0531.2632.3331.79
BERT_FC-GCN+sen81.7577.3979.5162.8555.7159.07
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, J.; Chen, Y.; Qin, Y.; Huang, R.; Zheng, Q. A Feature Combination-Based Graph Convolutional Neural Network Model for Relation Extraction. Symmetry 2021, 13, 1458. https://doi.org/10.3390/sym13081458

AMA Style

Xu J, Chen Y, Qin Y, Huang R, Zheng Q. A Feature Combination-Based Graph Convolutional Neural Network Model for Relation Extraction. Symmetry. 2021; 13(8):1458. https://doi.org/10.3390/sym13081458

Chicago/Turabian Style

Xu, Jinling, Yanping Chen, Yongbin Qin, Ruizhang Huang, and Qinghua Zheng. 2021. "A Feature Combination-Based Graph Convolutional Neural Network Model for Relation Extraction" Symmetry 13, no. 8: 1458. https://doi.org/10.3390/sym13081458

APA Style

Xu, J., Chen, Y., Qin, Y., Huang, R., & Zheng, Q. (2021). A Feature Combination-Based Graph Convolutional Neural Network Model for Relation Extraction. Symmetry, 13(8), 1458. https://doi.org/10.3390/sym13081458

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop