Next Article in Journal
Does Power Output at Critical Power Intensity Interchange between Cycling and Running?
Previous Article in Journal
Application of Fruit By-Products and Edible Film to Cookies: Antioxidant Activity and Concentration of Oxidized LDL Receptor in Women—A First Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Key News Event Detection and Event Context Using Graphic Convolution, Clustering, and Summarizing Methods

1
School of Information Engineering, Minzu University of China, Beijing 100081, China
2
School of Chinese Ethnic Minority Languages and Literatures, Minzu University of China, Beijing 100081, China
3
Key Laboratory of Ethnic Language Intelligent Analysis and Security Governance, Ministry of Education, Minzu University of China, Beijing 100081, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(9), 5510; https://doi.org/10.3390/app13095510
Submission received: 31 March 2023 / Revised: 12 April 2023 / Accepted: 24 April 2023 / Published: 28 April 2023
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
The detection of key events and identification of the events’ context have been widely studied to detect key events from large volumes of online news and identify trends in such events. In this paper, we propose a Key News Event Detection and Context Method based on graphic convolving, clustering, and summarizing methods. Our method has three main contributions: (1) We propose the use of position vectors as time-embedding feature representations and concatenate semantic and time-embedding features as node features of the graph to distinguish different nodes of the graph. Additionally, a temporal nonlinear function was constructed using time embedding to objectively describe the effect of time on the degree of association between nodes. (2) We update the graph nodes using a graph convolutional neural network to extract deep semantic information about individual nodes of a high-quality phrase graph, thereby improving the clustering capability of graph-based key event detection. (3) We apply a summary generation algorithm to a subset of news data for each key event. Lastly, we validated the effectiveness of our proposed method by applying it to the 2014 Ebola dataset. The experimental results indicate that our proposed method can effectively detect key events from news documents with high precision and completeness while naturally generating the event context of key events, as compared to EvMine and other existing methods.

1. Introduction

The aim of so-called key event detection is to obtain a set of clusters in a corpus about a certain topic, with each cluster containing a set of articles describing the same event. The event context is a process of generating event descriptions for each key event and arranging them by time of occurrence. With the rapid development of computer and network technology, an exponentially growing, complex, and massive amount of online data and information has emerged. As a result, people often only see partial information, and it is thus difficult to obtain a comprehensive and quick overview of the big picture. Key event detection can be involved in applications across different modalities, such as video summarization [1] and augmented reality [2]. Accordingly, the search for more advanced techniques to detect events and their contexts, enabling the rapid and accurate detection of key events in large-scale data and the timely observation of the context of such key events [3], has become a growing subject for many researchers.
News events evolve rapidly and are variable in nature [4]. It is therefore difficult to obtain large amounts of annotated data. Meanwhile, the models obtained on the basis of training data are less portable. Thus, the use of unsupervised methods such as clustering for the detection of key events has also become a trend. For example, Yang et al. [5] constructed a KeyGraph and then used a community discovery algorithm for clustering. A Burst Information Network (BINet) based event detection model, the Time-aware Hierarchical Bayesian Model (TaHBM), was proposed by Ge et al. [6]. Zhang et al. proposed an unsupervised key event detection model, EvMine [7]. Both of the above methods incorporate time information. Usually, a key event is discussed in articles with similar content and is published within a certain timeframe. After an important news event occurs, it is widely discussed in the following days, but the discussion gradually fades as the time frame expands. This shows that time information can serve as an indicative factor for describing different key events in news. Integrating time information into clustering algorithms can improve their clustering effectiveness. Typically, a news article includes not only the text of the news itself, but also the time of publication. This publishing pattern provides a natural convenience for utilizing time information. Among the aforementioned models, EvMine is a more effective method for event detection and has been successfully applied in detecting key events. However, EvMine has the following limitations: (1) The practice of artificially assigning weights to edges between graph nodes based on experience is highly subjective and lacks generalizability. (2) The deep semantic information contained in the nodes of the original peak phrase graph is not considered. The nodes are clustered in the original peak phrase graph, and the clustering results are often unsatisfactory. (3) As EvMine focuses on key event detection, it lacks automatic identification of the key event context. As a result, there is a need to improve the utilization of time information in constructing the connected edges of a graph to obtain high-quality semantic feature information for the nodes and enable automatic identification of the context of key events. These three key points are crucial for the study of key event detection and event context.
To address the above issues, inspired by the EvMine model, this paper proposes a method for detection of key events in news documents and identification of the events’ context based on graphic convolving, clustering, and summarizing (GCS). The work presented in this paper mainly includes: (1) The union set of the semantic embedding of high-quality phrases extracted from the news corpus and the embedded release times of the documents to which the phrases belong are used as the nodes of the graph. The nodes are connected based on the news release rules in reality, thereby constructing an initial graph of high-quality phrases. (2) A graph convolutional network (GCN) is applied to high-quality phrase graphs to obtain deep semantic features of each node. (3) The advanced high-quality phrase graph consists of the semantic relatedness between two different nodes and the weights of the edges updated with the temporal relatedness between the two nodes described by a nonlinear function. (4) Nodes of advanced high-quality phrase graphs are clustered using the Louvain community discovery algorithm to classify news documents into subsets of news data for different key events. (5) The summary generation algorithm is applied to each subset of key event news data to generate summary descriptions of key events. The summary descriptions are then automatically ranked based on the span of news releases in the key event news. Finally, the event context of the events is obtained.

2. Related Work

According to the format of input news text, the key event detection task can be divided into sentence-level event detection and document-level key event detection. For sentence-level event detection tasks, using graph structures has become a trend. Xie et al. [8] proposed a graph parsing method using a dependency-tree-based graph and convolutional neural networks. He et al. [9] presented a self-constructed dependency and graph convolution network for event detection. For document-level key event detection, the objective of certain recent studies is to identify narratives within a broad collection of news articles. Saravanakumar et al. [10] proposed a hierarchical clustering algorithm that leverages both textual and contextual features to group similar news stories together. Sia et al. [11] discussed the use of clusters of pre-trained word embeddings for fast and effective topic modeling. Santos et al. [12] presented a method for simplifying multilingual news clustering through projection from a shared space. The trend of using graph structures has also begun to become popular in document-level key event detection. Yang et al. [5] first constructed a KeyGraph based on textual keyword co-occurrence relationships, then combined this with a community detection algorithm to segment the KeyGraph and used the extracted topic features to detect events in social media streams. S. Gaglio et al. [13] performed frequent pattern mining based on keywords, using sets of words that frequently occur together as events. Ge et al. [6] proposed a Burst Information Network (BINet) based event detection model, the Time-aware Hierarchical Bayesian Model (TaHBM), which first identifies key nodes or key regions in a BINet as the center of mass and constructs clusters, thus transforming the event detection problem into a community detection implementation on a BINet network. Brochier et al. [14] proposed document network embedding with topic word attention. One of the most commonly used community discovery algorithms is the Louvain algorithm [15], which is known for its speed and scalability, making it particularly useful for large-scale networks. Zhang et al. [7] proposed an unsupervised key event detection model, EvMine. Furthermore, the integration of time information has also become a research direction in these methods. Ge et al. [6] incorporated time information into the Bayesian model, resulting in an improvement in performance. Zhang et al. [7] constructed a peak phrase graph through statistical analysis of time information, which improved the effectiveness of document clustering in subsequent stages.
Among these research methods, the selection of keywords is particularly important. Guille A et al. [16] proposed a classical keyword generation method. Chen et al. [17] proposed a topic model that uses the lexical relevance of words in a given dictionary to extract coherent keywords for a topic. Due to the broad applicability of phrases to text-related tasks, it is often used as a form of expression for keywords. Some researchers [18] used an N-Gram for event detection, but this also invokes some nonsensical collocations and requires specific methods to select semantic phrases. Gu et al. [19] proposed a new unsupervised context-aware quality phrase tagger, UCPhrase. In the unsupervised case, the context of each N-Gram was viewed to find the largest pattern and to be able to find more complete phrases that served as better supervised signals for the feature-based classifier for the UCPhrase context. A pre-trained language model with a transformer containing an attention mechanism is then used to identify high-quality phrases in new sentences.
However, the semantic feature extraction of keywords determines the accuracy of the clustering results. Microsoft Palangi et al. [20] proposed a text matching model based on a long short-term memory recurrent neural network (LSTM-RNN), which transforms query words and target documents into vectors and calculates cosine similarity. Zhang and Liu et al. [21] implemented a bi-directional LSTM by proposing a sentence-state-based LSTM. It maintains a global structured sentence-level semantic vector that is used and updated by each recurrent unit and can better capture dependencies based on global information. To more rationally describe the semantic features of long texts, by combining graph representations with neural networks, graph convolutional neural networks [22] were generated. Yao et al. [23] proposed the TextGCN model to learn word and document semantic representations and perform text classification without pretraining based on word co-occurrence matrices and graph convolution of heterogeneous semantic association graphs generated from document–word associations.

3. Methodology

Figure 1 shows the general structure of the GCS proposed in this document, which consists mainly of the following parts: (1) graph construction based on high-quality phrases; (2) graph update based on GCN; (3) graph-based key events detection; and (4) key event context generation. The content of each part is described in detail below.

3.1. Graph Construction Based on High-Quality Phrases

3.1.1. Semantic Embedding of High-Quality Phrases

Given a corpus D = d 1 , d 2 , , d n , where d i is the i th document, i 1 , n , and n is the number of documents in the corpus, a classifier f . ; θ is trained using UCPhrase to extract high-quality phrases from each document:
H P i = f d i ; θ
where H P i = h p 1 i , h p 2 i , is the set of high-quality phrases in the   i th document, and h p j i is the j th high-quality phrase in the i th document, h p j i h p k i , k j . The number of high-quality phrases per document is variable.
Suppose D H P = H P 1 , H P 2 , , H P n M is the set of high-quality phrases in the entire documents. M is the total number of high-quality phrases extracted:
D H P = i = 1 n H P i
Assume that h p j i high-quality phrase appears N p times in a different document, i.e., h p j i = h p k m = = h p g v N p , m i v . The semantic embedding of the high-quality phrase itself is denoted by s h p j i . The semantic embedding of high-quality phrase contexts is denoted by m h p j i . Based on this annotated form, the semantic embedding of the high-quality phrases h p k m and h p g v is denoted as s h p k m ,   m h p k m ,   s h p g v ,   m h p g v . Then the final semantic embedding e h p j i of h p j i is given by [7]:
e h p j i = 1 N p s h p j i , m h p j i + s h p k m , m h p k m + + s h p g v , m h p g v
Using Equation (3) to semantically embed each high-quality phrase in the set D H P , to obtain the set of semantic embeddings of high-quality phrases E D H P = E H P 1 , E H P 2 , , E H P n M :
E D H P = i = 1 n E H P i
where E H P i = e h p 1 i , e h p 2 i , is the set of semantic embeddings of high-quality phrases of the ith document.

3.1.2. Time Embedding of High-Quality Phrases

As the same phrases may appear at different times, temporal information needs to be used to distinguish these nodes at different points in time. In news articles, a key event is usually discussed in a number of articles that are similar in content and are published within a certain time frame. After a key news event has occurred, it is widely discussed in the days that follow, but after the time frame has been extended further, the discussion fades. Therefore, this paper proposes a time-embedding approach for news articles, which can be used to express the relative relationship between the publication times of articles. As temporal information is also simultaneously spatial, time can be understood using the concept of relative position of time. Inspired by positional coding [24], the use of positional information to encode time allows the expression of relative positional information of time.
Given that there are news articles containing high-quality phrases h p j i for k days in the corpus, and the vector corresponding to the publication time of a news document on a certain day is the d -dimensional time embedding t i n d e x , , i n d e x 1 , k , then the time embedding corresponding to the earliest release time is t 1 and the time embedding corresponding to the latest release time is t k . So, the time embedding is given by:
t i n d e x l h p j i = sin i n d e x 10000 l d ,                     if   l   is   even   cos i n d e x 10000 l 1 d ,                   if   l   is   odd  
where l is the l th element in the time-embedding vector, l     1 , d .

3.1.3. Graph Construction of High-Quality Phrases

Using Equations (3)–(5), the semantic embedding of each phrase in the set of high-quality phrase semantic embedding and the corresponding news release time embedding of the article in which the phrase is embedded are concatenated as the composite high-quality phrase semantic embedding features, i.e.,
te h p j i = e h p j i , t i n d e x l h p j i  
The final set of composite high-quality phrase semantic embedding T E D H P = T E H P 1 , T E H P 2 , , T E H P n M is formed:
T E D H P = i = 1 n T E H P i
where T E H P i = t e h p 1 i , t e h p 2 i , is the set of semantic embedding of compound high-quality phrases for the i th document.
Using each compound high-quality phrase semantic embedding in the T E D H P as a node feature of the graph, the constructed high-quality phrase graph should have M nodes.
According to the general pattern of news releases in reality: (1) The same key event is usually reported by several media during the same day. So, if the news documents to which the nodes belong have the same publication date, there is an edge between the nodes. (2) News describing a key event breaks on the first day and is reported in the following days. So, if the same phrase appears on consecutive days, there is an edge between these nodes. For node-to-node associations, the high-quality phrase graph is constructed as an undirected graph based on this general pattern of news releases. Meanwhile, if there are edges between two nodes, the initial weights of the edges are both set equal to 1. Figure 2 shows this.

3.2. Graph Optimization

3.2.1. GCN-Based Graph Node Optimization

The high-quality phrase graphs that are currently being constructed have nodes that only consider shallow semantic information and their representation ability is not strong. The graph nodes need to be updated by a graph convolutional neural network to extract deep semantic information. This operation facilitates the clustering or subsequent processing of the graph.
First the nodes of the high-quality phrase graph are denoted as n 1 , n 2 , , n M , where M is the number of nodes. The node number is matched with the semantic embedding of the compound high-quality phrase in     T E D H P = T E H P 1 , T E H P 2 , , T E H P n M , i.e., n i is the i th compound high-quality phrase semantic embedding, including the phrase embedding and the corresponding time-embedding feature t i . Let the high-quality phrase graph be G = N , E , where N = n 1 , n 2 , , n M is the set of nodes in G and E is the set of edges in G . When performing graph convolution on G = N , E , the time embedding in node n i is not considered, i.e., another graph G ^ = N ^ , E is constructed, where N ^ = n 1 t 1 , n 2 t 2 , , n M t m . A graph convolution update is performed on G ^ = N ^ , E to extract the deep semantic features of the nodes and obtain a high-quality phrase graph with deep semantic features. The update process is given by Equation (8) [22]:
G ^ l + 1 = R e l u D 1 2 A ˜ D 1 2 G ^ l W l
where, for a GCN with L layers, l 1 , L , G ^ l is the state at graph input, and G ^ l + 1 is the output at layer l . D is the degree matrix of the graph G ^ l , A ˜ = A + I is the adjacency matrix with added self-connections, and A is the adjacency matrix of G ^ l . I is the identity array. R e l u is the activation function, and W is a parameter in the graph convolutional neural network.
After the GCN convolution operation, a new high-quality phrase graph G = N ^ , E is obtained, where N ^ = n ^ 1 + t 1 , n ^ 2 + t 2 , , n ^ M + t M , and n ^ i denotes the deep semantic embedding feature of the ith node in the new high-quality phrase graph.

3.2.2. Optimizing Graph Edges

When constructing the initial high-quality phrase graph, the weights of the edges between the nodes are simply set to 1. In practice, however, the weights of the edges between the nodes are closely related to the following factors: (1) the cosine similarity of node-to-node phrase embedding and (2) the influence of time between nodes. A key event in a news article is usually discussed in large bursts by a number of articles, and the heat of that key event decreases more rapidly over time. However, the effect of time information on the nodes is not presented as a conventional linear relationship. Therefore, this paper constructs a nonlinear exponential function I u , v to simulate this law, as shown in Equation (9).
I u , v = e 1 + C o s S i m t u , t v
where C o s S i m denotes cosine similarity, u 1 , M , v 1 , M denotes two nodes in the graph, and t u , t v is the corresponding time embedding.
The weights of the edges are closely related to the cosine similarity of the phrase embedding between nodes, in addition to the influence of time between nodes. Therefore, taking these factors into account, the cosine similarity of phrase embedding between nodes and Equation (9) are combined to design the calculation of edge weights w u , v in high-quality phrase graphs, as shown in Equation (10).
w u , v = I u , v max C o s S i m n ^ u , n ^ v , 0
where n ^ u , n ^ v are the deep semantic embedding features corresponding to the two nodes u , v .

3.3. Key Event Detection

Graph-Based Community Discovery: Since the phrases are extracted from a massive amount of news text, their quantity is also quite large. As a result, the constructed graph structure is also relatively large. Based on the entire process, an unsupervised community detection algorithm needs to be selected. Additionally, it is hoped that the algorithm can perform calculations quickly. The Louvain community detection algorithm typically performs well when dealing with large networks and is an unsupervised algorithm that is computationally efficient. Therefore, this algorithm was chosen for this stage. The Louvain algorithm [15], a modularity-based community discovery algorithm, is used to detect non-overlapping community structures in graphs. The goal of optimization is to maximize the modularity of the entire structure of the graph (community network). Modularity is used to describe the density of communities in the graph structure. The algorithm initially treats each node as a community, and considers the neighboring nodes of each community separately, attempting to merge them. It finds the partition method with the largest change in modularity and divides it into the same community. Iterative calculations are performed until there is no further change. The modularity formula is given in Equation (11).
Q = 1 2 R u v w u v k u k v 2 R δ u , v
where u and v are two nodes in G , with u     1 , M and v     1 , M . R represents the total number of edges in G and w u v represents the weight of the edge between nodes u and v . k u and k v are the degree of nodes u and v , respectively. If nodes u and v are in the same community, δ u , v = 1 ; otherwise, δ u , v = 0 .
The algorithm has a low time complexity and is suitable for large-scale networks and works well. It is currently one of the most commonly used and efficient algorithms. Therefore, in this paper, this algorithm is used to cluster the deep semantic high-quality phrase graph G and initially generate the key event family C , as shown in Equation (12).
C = Louvain G
where C is the set of clusters generated using Louvain’s algorithm, C = C 1 , C 2 , , C r , and r is the total number of communities found. Clusters with node number 1 are omitted, leaving each cluster containing two or more nodes, denoted as C .
Extraction of Key Event Document: Based on Equation (12), the news documents in corpus D are divided into sets (or clusters) of key event documents corresponding to each cluster of deep sub-semantic high-quality phrases in C . However, using only phrases for matching may lose documents that express the same event but with different wording. The article selection module IKEDS [7], based on the results of community discovery, enables the selection of documents for the same event but using different wording. This selection module, IKEDS, is used in this paper to further process the community discovery result C of the deep high-quality phrase graph in order to extract the complete (identically and differently worded) key event documents corresponding to each cluster of deep semantic high-quality phrases in C , denoted by K i , which represents the set of all documents for the ith key event extracted. Based on the above method, Equation (13) is constructed to extract the complete set of key event documents corresponding to C , denoted by K .
K = IKEDS C
where K = K 1 , K 2 , , K l , K i D . l   represents the number of key events.

3.4. Generation of Key Events Context

News events are important events in society that attract attention, so there is a surge in news coverage after a news event has occurred. Simply reading the news is inefficient for objectives such as opinion analysis, which aims to understand the context and trends of news events. Therefore, it is important to generate a news context based on the news text of the key event and an estimate of the segment in which it occurred, after the key event has been detected. This involves two tasks: generating a summary description of each key event and obtaining an estimate of the time span in which each key event occurred.

3.4.1. Abstract Generation

The key event detection task aims to detect key events from the corpus on the same topic, i.e., a set of thematically coherent documents on a specific topic. However, for these key events, it is not possible to visualize the specific description of the key events from the text. For this reason, in this paper, the top-k articles from the document clusters under each key event are selected, and the descriptions of the key events are automatically generated based on a multi-text summary generation method.
In this paper, we use the unsupervised extractive summary generation method Text-Rank [25]. This method does not require the construction of a summary tag for the text, is simple to deploy, and works well without the limitation of the length of the input text. Therefore, this method is used for the generation of summaries. The method first extracts key sentences, then constructs a graph of similarity between sentences, and then calculates the ranking of the sentences according to the similarity. The formula used in this paper to calculate the similarity between two sentences is as follows:
S i m i l a r i t y S i , S j = C o s S i m k = 1 c o u n t S i w k i c o u n t S i , k = 1 c o u n t S j w k j c o u n t S j
where S i , S j are any two sentences in the text, C o s S i m . is the cosine similarity calculation function, c o u n t . is used to calculate the number of words in the sentence, and w k i is the word vector of the k th word in the ith sentence. The average of the word vectors of all the words in the sentence is used as the representation of the vector of the sentence.
After using the Text-Rank algorithm, the sentences in a multi-document document are sorted and the top-n sentences are finally selected as summary. Let l key events K = K 1 , K 2 , , K l be predicted, and the top-k news articles in each key event K i are used to generate summary S i as the description of key event K i using Text-Rank.

3.4.2. Identifying Event Context

When identifying the event context, in addition to the textual descriptions of each key event, the generation of a timeline for each key event is also required. The events are arranged in chronological order so that the sequence of events can be determined. In the section above, we have obtained a description of each key event and the description of that event. Therefore, here it is necessary to obtain an estimate of the time of occurrence of each key event from the corpus.
In the news corpus, each news item contains its publication time   t . In a key event K i , there is a series of related news articles in which there is an earliest news release time t s t a r t , and a latest news release time t e n d . Thus, the time span T i = t s t a r t   , t e n d of a key event can be estimated based on the publication time of each article in a key event. The pseudocode of the process is shown as Algorithm 1.
Algorithm 1 Extract Time Estimation
Input: Key Event K
Output: Time Estimation T
1. t_start ← MAX_TIME
2. t_end ← MIN_TIME
3. for d ← K do
4.  publish_time ← d.publish_time
5.  if publish_time < t_start do
6.    t_start ← publish_time
7.  end if
8.   if publish_time > t_end do
9.    t_end ← publish_time
10.  end if
11. end for
12. T ←(t_start, t_end)
13. return T
Based on the summary S i of each key event document cluster and the time estimate T i , the event context context S L can be obtained, as shown in the following equation:
S L = i = 1 l { S i } + T i
where S L = s l , s l 2 , , s l l , S L i = T i , S i , where l is the number of predicted key events. Ranking SL according to the start time t s t a r t in the time span estimate T i , the final event context pulse result S L is obtained.

4. Experiments

In this section, selected baseline methods for key event detection are compared with the method proposed in this paper in terms of metrics such as precision, recall, and f-measure. Ablation experiments were conducted to verify the effectiveness of the deep semantic information of the nodes in the graph extracted by the GCN proposed in this paper. It also identifies the key event context.

4.1. Experiment Setup

This experiment was conducted on a machine with 128 GB memory, a CPU model of Intel(R) Xeon(R) CPU E5-2620 v4, and a GPU model of NVIDIA GeForce GTX 1080 Ti. The operating system uses Ubuntu 9.4.0, the programming language uses Python 3.7, the deep learning platform uses PyTorch, and PyCharm is used as the integrated context environment.
Experiments were conducted using the Ebola dataset [12], which includes a large number of news articles on the ‘ebola outbreak 2014’ published from 18 September 2014 to 31 October 2014, each with its time of publication. It contains 741 news articles and 17 key events.

4.2. Metrics

Three evaluation metrics commonly used in previous key event detection studies [7] were used: precision, recall, and f-measure. Precision was used to assess the independence of the events detected by each cluster. Recall was used to assess the detected completeness of the key events. F-measure is the harmonic mean of precision and recall.
Specifically, let the predicted l document clusters be K = K 1 , K 2 , , K l ; the subsequent calculation needs to consider the top-5 articles in the document cluster K i , so the number of articles in each cluster needs to be >5, where each K i is a set of articles ranked according to the key event matching degree, and the higher the ranking indicates the higher the relevance to the key event. Assuming that there are N   real key events in D , the existence of N real key events is denoted as G = G 1 , G 2 , , G N . If more than half of the top-five articles in K j belong to G i , then K j and G i are considered to match each other, noted as M a t c h G i , K j = 1 , and M a t c h G i , K j = 0 if there is no match. The equations for calculating precision and recall are defined below.
c h e c k G i 1 ,   i f     C i   l e t   M a t c h G i , K j = 1 0 ,   e l s e
p r e c i s i o n = G i c h e c k K i l
r e c a l l = G i c h e c k K i N
F-measure is the harmonic mean of precision and recall.

4.3. Experimental Results and Analysis

The following key event detection methods were selected for comparison:
  • Miranda et al. [26] can classify emerging documents into existing document clusters by training an SVM classifier.
  • newsLens [27] clusters documents by processing several overlapped time windows.
  • Staykovski et al. [28], which is a modification of newsLens.
  • S-BERT [29] uses Sentence Transformers to obtain a vector representation of documents, which is then clustered by processing a time window.
  • EvMine [7] clusters documents using graphs constructed from detected bursts of temporal peak phrases within a certain time range.
To avoid contingency, the results of each method were obtained by averaging the results of five experiments, which are shown in Table 1.
As can be seen from Table 1, the proposed method outperforms the other benchmark methods in terms of precision and F-measure on the Ebola dataset. The proposed method outperforms the other benchmark methods by 6 to 13 percentage points in these two metrics. The high independence of the events detected by the model in each cluster can be seen in the precision. The score of F-measure demonstrates the good ability of our method to handle noisy data. Some of the baseline methods produce high recall values, which, after analysis, is due to the fact that these methods produce small but many key events and, therefore, coverage of key events can be high. However, since they also generate many of the same key events, their precision values are lower.

4.4. Ablation Study

In this paper, we primarily propose methods for representing and utilizing time information, as well as extracting deep semantic information using a GCN. To verify the effectiveness of these two methods proposed in this paper, we designed and conducted two different ablation experiments.
The first module we chose to verify is the module for representing and utilizing time information. After removing the GCN updating module, the primary difference between our proposed method and Zhang et al.’s EvMine [12] method is the approach for expressing and utilizing time information. EvMine utilizes temporal information to filter phrases and incorporate time information into the method. In contrast, our proposed method uses the concept of position to represent temporal information. As shown in Table 2, the comparative results are as follows. From the table, it can be seen that our proposed method, after removing the GCN module, outperforms EvMine in both precision and F-measure. This demonstrates the effectiveness of our proposed method for representing and utilizing time information.
The second module we chose to verify is the module for utilizing the GCN to extract deep semantic information. We compared the performance with and without the GCN module. When removing the GCN module, only the semantic features of the original phrases were used for clustering without utilizing the deep semantic information extracted by the GCN combined with graph structure. The comparative results are shown in Table 3. The results show that the accuracy, recall, and F-measure are significantly improved when the GCN module is added. This fully demonstrates the effectiveness of the proposed module for utilizing the GCN to extract deep semantic information.

4.5. Identifying the Event Context

As the detected key events are a set of topic-coherent documents on a specific topic, it is not possible to literally visualize the description of a specific key event. Therefore, an unsupervised extractive summary generation algorithm is used to extract a summary S i of the top-k articles for each key event to obtain a textual description of the key event. This number of articles k can be freely formulated. Here, k = 10. S i is taken to be combined with the time estimate T i obtained for each key event in Section 3.4. Finally, the event context   S L = S L 1 , S L 2 , , S L l , where l is the number of predicted key events and S L i = T i , S i . For the Ebola dataset, the generated event context is shown in Figure 3.

4.6. Parameter Study

In this section, the impact of the parameter γ on the performance of our proposed method is analyzed. γ is the number of layers of the GCN. Since graph convolutional neural networks can be stacked with layers in the model, the difference in the number of layers will inevitably also affect the performance of the method, so experiments were conducted on the GCN with different numbers of layers. Different numbers of layers from one to five were considered. To avoid chance, all five experiments were performed and averaged, as shown in Figure 4.
We observed that in the initial stages, the method’s performance improved as the number of GCN layers increased. This is because the semantic information extracted becomes more enriched with an increasing number of layers. The method achieved its best performance with two to three GCN layers. However, beyond three layers, the method’s performance started to degrade. This is due to the information becoming overly smoothed, which leads to a loss of semantic expression ability. Therefore, excessive stacking of layers should be avoided, and setting the parameter to two or three is optimal.

4.7. Discussion

In this section, we discuss the advances and limitations of the proposed methodology.
The first aspect we would like to discuss is the advance of our proposed method, which showed significant improvements in accuracy and F-measure values compared to other baseline methods. These results are further supported by ablative experiments, which validate the effectiveness of time-related features, as well as the proposed GCN update module.
Moving on to the limitations of our proposed method, we should note that due to its increased computational complexity, our method takes about 14% longer to process the same dataset compared to similar methods when tested on our experimental platform. It is worth noting, however, that our proposed method achieved significant improvements in evaluation metrics such as accuracy.

5. Conclusions and Future Work

In this paper, a key event detection and event context pulse grooming method is proposed. A method is proposed for the embedding feature representation of the temporal relative position. A temporal non-linear function is also constructed to describe the effect of time on the degree of association between nodes. The subjective use of temporal information in previous studies is avoided. The nodes of the graph are updated using a graph convolutional neural network to extract deep semantic information about individual nodes of a high-quality phrasal graph, thereby enhancing the graph’s expressive ability. The summary generation algorithm is applied to each subset of key event news data, which is then sorted according to time span to obtain the final event context, thus avoiding the need to manually generate the relevant descriptions of individual key events. Experiments on the Ebola corpus demonstrate the effectiveness of our proposed approach, achieving higher F-measure and precision values compared to existing baselines methods. There is a little research on this task and the proposed method offers some new ideas and practical implications for current key event detection tasks. For example, it can be used to detect key events in the fields of disease epidemics, natural disasters, etc.
Considering that the use of the original GCN alone may not be sufficient to extract the deep semantic features of the nodes in the graph, future work will try to use more complex graph networks for extraction. This paper currently only performs key event detection on news documents under the same top-level topic, and future work will continue to explore the task of simultaneous key event detection from a corpus containing events from multiple top-level topics.

Author Contributions

Z.L. proposed overarching research goals and aims. Y.Z. and Y.L. performed the experiment. Y.Z. and Y.L. wrote the main manuscript text. C. proposed the research framework. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2020YFB1406702-3, in part by the National Natural Science Foundation of China under Grant 62006257.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used for this work is the Ebola News dataset published by Zhang [7]. It can be found at https://github.com/yzhan238/EvMine/tree/main/data (accessed on 1 January 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hussain, T.; Muhammad, K.; Ullah, A.; Cao, Z.; Baik, S.W.; de Albuquerque, V.H.C. Cloud-assisted multiview video summarization using CNN and bidirectional LSTM. IEEE Trans. Ind. Inform. 2019, 16, 77–86. [Google Scholar] [CrossRef]
  2. Minaee, S.; Liang, X.; Yan, S. Modern Augmented Reality: Applications, Trends, and Future Directions. arXiv 2022, arXiv:2202.09450. [Google Scholar]
  3. Goyal, P.; Kaushik, P.; Gupta, P.; Vashisth, D.; Agarwal, S.; Goyal, N. Multilevel Event Detection, Storyline Generation, and Summarization for Tweet Streams. IEEE Trans. Comput. Soc. Syst. 2020, 7, 8–23. [Google Scholar] [CrossRef]
  4. Salwen, Z.C.; Michael, B.; Garrison, B.; Driscoll, P.D. (Eds.) Communication Research Trends. In Online News and the Public; Routledge: London, UK, 2006; Volume 25, pp. 37–39. [Google Scholar]
  5. Yang, S.; Sun, Q.; Zhou, H.; Gong, Z.; Zhou, Y.; Huang, J. A topic detection method based on Key Graph and community partition. In Proceedings of the 2018 International Conference on Computing and Artificial Intelligence, Shenzhen, China, 8–10 December 2018; ACM: New York, NY, USA, 2018; pp. 30–34. [Google Scholar]
  6. Ge, T.; Pei, W.; Ji, H.; Li, S.; Chang, B.; Sui, Z. Bring you to the past: Automatic Generation of Topically Relevant Event Chronicles. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015; Volume 1, pp. 575–585. [Google Scholar]
  7. Zhang, Y.; Guo, F.; Shen, J.; Han, J. Unsupervised Key Event Detection from Massive Text Corpora. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD’22), Washington, DC, USA, 14–18 August 2022; Association for Computing Machinery: New York, NY, USA, 2022; pp. 2535–2544. [Google Scholar] [CrossRef]
  8. Xie, J.; Sun, H.; Zhou, J.; Qu, W.; Dai, X. Event Detection as Graph Parsing. In Findings of the Association for Computational Linguistics: ACL-IJCNLP, Online; Association for Computational Linguistics: Toronto, ON, Canada, 2021; pp. 1630–1640. [Google Scholar] [CrossRef]
  9. He, L.; Meng, Q.; Zhang, Q.; Duan, J.; Wang, H. Event Detection Using a Self-Constructed Dependency and Graph Convolution Network. Appl. Sci. 2023, 13, 3919. [Google Scholar] [CrossRef]
  10. Saravanakumar, K.K.; Ballesteros, M.; Chandrasekaran, M.K.; McKeown, K. Event-Driven News Stream Clustering using Entity-Aware Contextual Embeddings. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Online, 19–23 April 2021; Association for Computational Linguistics: Toronto, ON, Canada, 2021; pp. 2330–2340. [Google Scholar] [CrossRef]
  11. Sia, S.; Dalmia, A.; Mielke, S.J. Tired of Topic Models? Clusters of Pretrained Word Embeddings Make for Fast and Good Topics too! In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; Association for Computational Linguistics: Toronto, ON, Canada, 2020; pp. 1728–1736. [Google Scholar] [CrossRef]
  12. Santos, J.; Mendes, A.; Miranda, S. Simplifying Multilingual News Clustering Through Projection From a Shared Space. arXiv 2022, arXiv:2204.13418. [Google Scholar] [CrossRef]
  13. Gaglio, S.; Re, G.L.; Moranam, M. A framework for real-time Twitter data analysis. Comput. Commun. 2016, 73, 236–242. [Google Scholar] [CrossRef]
  14. Brochier, R.; Guille, A.; Velcin, J. Inductive document network embedding with topic word attention. In Proceedings of the Advances in Information Retrieval: 42nd European Conference on IR Research, ECIR 2020, Lisbon, Portugal, 14–17 April 2020; Proceedings, Part I 42. Springer: New York, NY, USA, 2020; pp. 326–340. [Google Scholar]
  15. Blondel, V.D.; Guillaume, J.-L.; Lambiotte, R.; Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech. Theory Exp. 2008, 10, 10008. [Google Scholar] [CrossRef]
  16. Guille, A.; Favre, C. Event detection, tracking, and visualization in Twitter: A mention-anomaly-based approach. Soc. Netw. Anal. Min. 2015, 5, 18–32. [Google Scholar] [CrossRef]
  17. Chen, Z.; Mukherjee, A.; Liu, B.; Hsu, M.; Castellanos, M.; Ghosh, R. Discovering Coherent Topics Using General Knowledge. In Proceedings of the 22st ACM International Conference on Information and Knowledge Management, San Francisco, CA, USA, 27 October–1 November 2013; ACM: New York, NY, USA, 2013; pp. 209–218. [Google Scholar]
  18. Guzman, J.; Poblete, B. Online relevant anomaly detection in the Twitter stream an eficient bursty keyword delection mode. In Proceedings of the ACM SIGKDD Workshop on Outlier Detection and Description, Chicago, IL, USA, 11 August 2013; ACM: New York, NY, USA, 2013; pp. 31–39. [Google Scholar]
  19. Gu, X.; Wang, Z.; Bi, Z.; Meng, Y.; Liu, L.; Han, J.; Shang, J. Ucphrase: Unsupervised context-aware quality phrase tagging. arXiv 2021, arXiv:2105.14078. [Google Scholar] [CrossRef]
  20. Palangi, H.; Deng, L.; Shen, Y.; Gao, J.; He, X.; Chen, J.; Song, X.; Ward, R. Deep sentence embedding using long short termmemory networks: Analysis and application to information retrieval. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 696–707. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Liu, Q.; Song, L. Sentence-State LSTM for Text Representation. arXiv 2018, arXiv:1805.02474. [Google Scholar]
  22. Thomas, K.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  23. Yao, L.; Mao, C.; Luo, Y. Graph convolutional networks for text classification. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 7370–7377. [Google Scholar]
  24. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 6000–6010. [Google Scholar]
  25. Rada, M.; Tarau, P. Textrank: Bringing order into text. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 15 January 2004; pp. 404–411. [Google Scholar]
  26. Miranda, S.; Znotins, A.; Cohen, S.B.; Barzdins, G. Multilingual Clustering of Streaming News. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 4535–4544. [Google Scholar]
  27. Laban, P.; Hearst, M.A. newsLens: Building and visualizing long-ranging news stories. In Proceedings of the Events and Stories in the News Workshop, Vancouver, BC, Canada, 4 August 2017; pp. 1–9. [Google Scholar]
  28. Staykovski, T.; Barrón-Cedeño, A.; Da San Martino, G.; Nakov, P. Dense vs. Sparse Representations for News Stream Clustering. In Text2Story@ ECIR 2019; ACM: New York, NY, USA, 2019; pp. 47–52. [Google Scholar]
  29. Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China, 3–7 November 2019; pp. 3982–3992. [Google Scholar]
Figure 1. The overall architecture of Key Event Detection from news documents and Event Context.
Figure 1. The overall architecture of Key Event Detection from news documents and Event Context.
Applsci 13 05510 g001
Figure 2. High-quality phrase graph. The time in the figure represents the publication time of the news document where the phrase was found, and the time range is determined by the corpus used, from the earliest published news in the corpus to the latest.
Figure 2. High-quality phrase graph. The time in the figure represents the publication time of the news document where the phrase was found, and the time range is determined by the corpus used, from the earliest published news in the corpus to the latest.
Applsci 13 05510 g002
Figure 3. The result of the key event context.
Figure 3. The result of the key event context.
Applsci 13 05510 g003
Figure 4. Impact of parameter γ .
Figure 4. Impact of parameter γ .
Applsci 13 05510 g004
Table 1. Key event detection results.
Table 1. Key event detection results.
PrecRecFmeas
Miranda et al. [26]0.4440.7060.615
newsLens [27]0.4260.8240.561
Staykovski et al. [28]0.4140.7060.522
S-BERT [29]0.5080.8360.631
EvMine [7]0.8290.6820.748
GCS0.9860.7060.824
Table 2. Comparison of the way time information is processed.
Table 2. Comparison of the way time information is processed.
PrecRecFmeas
EvMine [7]0.8290.6820.748
GCS-NoGCN0.9100.6470.758
Table 3. Proving the validity of the GCN module.
Table 3. Proving the validity of the GCN module.
PrecRecFmeas
GCS-NoGCN0.9100.6470.748
GCS0.9860.7060.824
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Z.; Zhang, Y.; Li, Y.; Chaomurilige. Key News Event Detection and Event Context Using Graphic Convolution, Clustering, and Summarizing Methods. Appl. Sci. 2023, 13, 5510. https://doi.org/10.3390/app13095510

AMA Style

Liu Z, Zhang Y, Li Y, Chaomurilige. Key News Event Detection and Event Context Using Graphic Convolution, Clustering, and Summarizing Methods. Applied Sciences. 2023; 13(9):5510. https://doi.org/10.3390/app13095510

Chicago/Turabian Style

Liu, Zheng, Yu Zhang, Yimeng Li, and Chaomurilige. 2023. "Key News Event Detection and Event Context Using Graphic Convolution, Clustering, and Summarizing Methods" Applied Sciences 13, no. 9: 5510. https://doi.org/10.3390/app13095510

APA Style

Liu, Z., Zhang, Y., Li, Y., & Chaomurilige. (2023). Key News Event Detection and Event Context Using Graphic Convolution, Clustering, and Summarizing Methods. Applied Sciences, 13(9), 5510. https://doi.org/10.3390/app13095510

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop