Next Article in Journal
Quantitative Evaluation of the View of the Landscape Using a Visibility Analysis Optimization Algorithm
Previous Article in Journal
Machine Learning Prediction Techniques in the Optimization of Diagnostic Laboratories’ Network Operations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multistage Mixed-Attention Unsupervised Keyword Extraction for Summary Generation

School of Information and Electronic Engineering, Hebei University of Engineering, No. 19 Taiji Road, Handan 056000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(6), 2435; https://doi.org/10.3390/app14062435
Submission received: 9 February 2024 / Revised: 1 March 2024 / Accepted: 10 March 2024 / Published: 13 March 2024

Abstract

:
Summary generation is an important research direction in natural language processing. Aimed at the problems of redundant information processing difficulties and an inability to generate high-quality summaries from long text in existing summary generation models, BART is the backbone model, an N + 1 coarse–fine-grained multistage summary generation framework is constructed, and a multistage mixed-attention unsupervised keyword extraction summary generation model is proposed (multistage mixed-attention unsupervised keyword extraction for summary generation, MSMAUKE-SummN). In the N-coarse-grained summary generation stages, a sentence filtering layer (PureText) is constructed to remove redundant information in long text. A mixed-attention unsupervised approach is used to iteratively extract keywords, assisting summary inference and enriching the global semantic information of coarse-grained summaries. In the 1-fine-grained summary generation stage, a self-attentive keyword selection module (KeywordSelect) is designed to obtain keywords with higher weights and enhance the local semantic representation of fine-grained summaries. Tandem N-coarse-grained and 1-fine-grained summary generation stages are used to obtain long text summaries through a multistage generation approach. The experimental results show that the model improves the ROUGE-1, ROUGE-2, and ROUGE-L metrics by a minimum of 0.75%, 1.48%, and 1.25% over the HMNET, TextRank, HAT-BART, DDAMS, and SummN models on summarized datasets such as AMI, ICSI, and QMSum.

1. Introduction

With the continuous development of large language models, text generation becomes a research hotspot in the field of natural language processing. It has gained a wide range of applications for tasks such as article creation, dialog generation, and conference summary. Summary generation is a task of text generation [1], and researchers use different techniques to generate summaries, including rewriting or extracting various information from documents [2]. Currently, lots of research results have been achieved in generating summaries for short texts [3,4]. However, it is still difficult to quickly obtain high-quality summaries from long textual data such as long documents, news interviews, and conference content. Therefore, improving the quality of summaries of long texts is of great research importance [5].
In the task of summary generation for long text, the current summary generation model has the problem of failing to comprehensively extract the important information of long text, and at the same time, there is redundant information in long text. Therefore, in order to better improve the generative ability of the summarization model, sentence filtering is usually used to remove redundant information in long texts during data preprocessing. To address the lack of an ability of summarization models to generate important information, based on deep learning models, supervised methods are designed by researchers to extract keywords, which require a large labeled training dataset. On the other hand, unsupervised methods do not require labeled datasets and have lower dataset requirements [6]. Therefore, unsupervised keyword extraction has been adopted by many scholars, and unsupervised keyword extraction methods based on text embedding, document embedding, and hybrid statistical embedding have emerged, such as MDERank [7], SIFRank [8], and KeyGames [9]. Although the above methods improve the extraction of keywords, the lower-weighted keywords may be used first for summarization inference, while the higher-weighted keywords fail to be applied, resulting in poor quality and readability of the summaries [10].
Therefore, the self-attention mechanism has been used by many scholars to select keywords with higher weights. For example, when generating summaries for long texts or multiple documents, a self-attention mechanism is employed by the Transformer model [11] to capture the dependencies between keywords in the input texts or documents. It achieves a focus on some of the important keywords. However, when the extracted keywords receive a small attentional weight, it affects the complete encoding of the keywords as well as the representation of important information.
The rest of this paper is organized as follows: Section 2 reviews the work related to this study. In Section 3, the framework of the summary generation model MSMAUKE-SummN is given and the components of the framework are described in detail. Experimental results and analysis related to the MSMAUKE-SummN model are shown in Section 4. Section 5 summarizes the conclusions of this paper and future research directions.

2. Related Work

Currently, Transformer-based summary generation models have made great progress in text summarization [11]. However, due to the large number of words in long texts and the sparse distribution of important information [12], the quality of summarization in long texts is poor as it is difficult for the summary generation model to capture the dependencies between important information in the text. A common solution is to reduce the input long text to short text; the BART model was proposed by Lewis et al. [13], which uses truncation of the input to reduce long text. Another approach is to improve the attention to cope with long text summarization tasks; Local Sensitive Hashing (LSH) attention was utilized by Kitaev et al. [14] to optimize the attention mechanism in Transformer to reduce the effect of quadratic complexity. In addition, a hierarchical self-attention mechanism was used by HMNet [15] and HAT-BART [16] to improve the input constraints of typical self-attention models, thus adapting the models to longer text inputs. Although the above methods can handle text content of a certain length, they are not flexible enough when the length of the text varies greatly.
The multistage modeling framework is capable of adapting to texts of different lengths by varying the number of stages, which is a more flexible way of processing compared to the truncated input and improved self-attention approaches. Therefore, the multistage summary generation model framework has been adopted by many scholars to adequately process information such as long texts or documents. A multistage summary generation framework structure, SummN, was proposed by Zhang et al. [17]. It is used to extract summaries of long texts and documents, and the final summary is obtained by N-coarse-grained and one-fine-grained summary generation stages. A staged long-text summary generation approach (PLSGA) was proposed by Fang et al. [18] to obtain the final summary. It includes key information extraction, extraction model training, generative model training, and candidate summary screening stages. A two-stage automatic summary generation model combining dual topic embedding and sentence absolute position embedding was proposed by Ren et al. [19]. It incorporates semantic features by introducing topic embedding in each of the two stages, extracts important information of the text in the extraction stage, and reduces the redundancy of the summary content. In addition, a sentence filtering layer was constructed by Mei et al. [20] for data preprocessing. It removes redundant sentences in long texts, shortens the text length, and provides the model with more streamlined data samples. The above methods reduce the adverse effects of excessively long text by moderating the number of stages, but these methods do not comprehensively consider the keywords and local semantic information of the text, resulting in summaries that omit important information.
The summary generation task needs to consider the importance of keywords for text summarization in addition to the effect of the length of the text. From past research work, it can be observed that with the continuous development of Pretrained Language Models (PLMs), many text generation studies have been centered around the technique of keyword extraction. Therefore, PLMs are used by many researchers as an embedding layer of keyword extraction models [21,22] to improve the extraction of keywords. Keyword extraction models are mainly classified into two categories, supervised and unsupervised. The use of supervised keyword extraction requires a large amount of labeled training data and the models generalize poorly outside the training data domain. Therefore, unsupervised learning was used by Eirini et al. [23] for keyword extraction, where the adjacency matrix corresponding to the word map of a target document is used as a vector representation of the vocabulary of that document, and distributed modeling is performed. To address the inability of embedding-based keyword extraction models to fully utilize PLMs, an unsupervised keyword extraction method, PromptRank, was proposed by Kong et al. [24]. It calculates the probability of keyword generation by the decoder based on prompts. Interpretable neural networks were used by Joshi et al. [25] for unsupervised keyword extraction, defining keywords as iconic words that are favorable for predicting the topic of a document. The above unsupervised approach simplifies the process of keyword extraction, but does not fully consider the importance of the keywords and the semantic relevance of the keywords to the sentences in the document. A mixed-attention model, AttentionRank, was proposed by Ding et al. [26] to extract keywords from documents in an unsupervised manner. It uses self-attention to determine the importance of keywords in the context of a sentence and cross-attention to calculate the semantic relevance between keywords and sentences in a document. The above methods were able to adequately extract the keywords from the text, but failed to enable the keywords to be used correctly in the inference of the summaries, resulting in the absence of important information in the summaries.
The uneven distribution of the extracted keywords in the original text and the lack of a rational method for their use lead to the fact that some keywords with lower weights may be used first in the inference of the summary, affecting the representation and distribution of important information in the summary. Therefore, in order to increase the detailed information of the summary, the targeted selection of keywords with higher weights to aid in inference is a critical step in summary generation. A method of selecting the best keywords was proposed by Tim et al. [27] to iteratively select keywords through an attention scoring function to obtain the best results overall. Meanwhile, in order to select more representative keywords, a probabilistic method was proposed by Akash et al. [28] for selecting representative keywords in a professional domain. A binary mixture model is used to obtain the distribution of keywords and select a representative subset of keywords. A keyword selection method was proposed by Venkatesh et al. [29]. The self-attentive head of the last layer of the pretraining encoder is clustered to select the most popular keywords from the original article. To address the problem of data sparsity in low-resource complex Named Entity Recognition (NER), an attention graph-aware keyword selection method was proposed by Ghosh et al. [30]. It uses attention graph-assisted selective masking to retain named entities in the input sentence and select contextually relevant keywords. The above methods select the most representative keywords in the text, and although they enhance the local semantic representation of the summary, there is some repetition in the use of keywords.
Aimed at the problems of high redundancy and poor comprehensiveness, accuracy, and readability in existing long text summaries, an N + 1 coarse–fine-grained multistage summary generation framework is constructed and a multistage mixed-attention unsupervised keyword extraction summary generation model (MSMAUKE-SummN) is proposed. During data preprocessing, the sentence filtering layer, PureText, is set to remove useless information from the data samples to obtain high-quality data samples. A mixture of self-attention and cross-attention is used to consider the importance and semantic relevance of keywords at the sentence level and document level. Multistage iterative extraction of keywords is used to facilitate summarization inference to enrich the global semantic information of coarse-grained summaries, rationalize the use of keywords using the self-attention mechanism, select more weighted keywords to assist the summary, and increase the detailed information of the summary to enhance the local semantic representation of the fine-grained summary.

3. MSMAUKE-SummN Model

Aimed at the problems of information redundancy, poor accuracy, comprehensiveness, and readability in current long-text summaries, an N + 1 coarse–fine-grained multistage summary generation framework is constructed, and a multistage mixed-attention unsupervised keyword extraction summary generation model (MSMAUKE-SummN) is proposed. The model framework contains two parts: the N-coarse-grained summary generation stage and the 1-fine-grained summary generation stage. In the N-coarse-grained summary generation stage, the sentence filtering layer, PureText, is constructed to remove redundant information from the original text (D) and the target text (T) to provide high-quality data samples for the summary generation model. An unsupervised approach with mixed attention is used to iteratively extract keywords to assist summary inference and enrich coarse-grained summary (Z) global semantic information. In the 1-fine-grained summary generation stage, self-attention is utilized to select keywords with higher weights to enhance the local semantic representation of the fine-grained summary (F). The framework of the MSMAUKE-SummN model is shown in Figure 1.
In Figure 1, an N + 1 coarse–fine-grained multistage summary generation framework is constructed, which connects N-coarse-grained and 1-fine-grained summary generation stages in series to obtain summaries of long texts in a multistage summary generation manner. In this case, in the N-coarse-grained summary generation stage, data filtering and segmentation, and coarse-grained summary generation are iterated N times. With the keyword-enhanced coarse-grained summary (Z) and the target text (T), the model is finetuned in the 1-fine-grained summary generation stage to obtain the final fine-grained summary (F). Specifically, in the process of data filtering and segmentation, the text filtering layer, PureText, is used to remove redundant information from the text, and the original text and the target text are segmented and paired using the ROUGE-based greedy algorithm. In the process of coarse-grained summary generation, mixed-attention screening of the initially extracted keywords is performed; the keywords at the sentence level and document level levels are paid full attention. The contextual relevance of the keywords to the sentences is considered. The final score of the keywords is calculated by linear combination to solve the problem of keyword duplication, so that more valuable keywords can assist the summary inference and enhance the global semantic representation of the summary. During fine-grained summary generation, self-attention focuses on selecting keywords with higher weights to aid summary inference and enhance the local semantic representation of the summary.

3.1. N-Coarse-Grained Summary Generation Stage

3.1.1. Data Filtering and Segmentation

In order to remove redundant information in long text, weakly supervised learning is used to train the BERT model in the N-coarse-grained summary generation stage. A versatile and lightweight sentence filtering layer (PureText) is constructed to filter out redundant sentences by utilizing text similarity. The process is to compare the ROUGE score of each sentence with the score of the standard summary sentence and then rank the importance of the sentence, filtering out up to 60% of the sentences in the text. By fine-tuning the BERT model, the ROUGE-1 metric value is utilized to distinguish the importance of different sentences. Since sentences are subunits of text with grammatical structure, it is natural to choose to filter sentences to produce a more concise text. It is hypothesized that the ROUGE-1 of a sentence is an important indicator that influences the evaluation of a summary. This was carried out by calculating the similarity scores between the sentences and the ROUGE-1 metric values of the standard summary sentences, labeling sentences with scores higher than the median as ‘important’ and sentences with scores lower than the median as ‘unimportant’.
After each sentence is labeled by the classifier, the sentences in the text are ranked according to their respective ‘importance’. The reward, R i , is defined according to the importance of the sentence. In the background of the 0–1 Knapsack algorithm, the set of sentences with the highest accumulated reward, R i , is found, while the length of the set of sentences does not exceed the input length limit (L) of the summary generation model. Each sentence is then weighted according to its number of tokens. Finally, the best set of sentences (filtered text) is obtained by filtering. In addition, the filtered original text (D) and target text (T) are segmented and matched to obtain the dataset ( D i , T i ) . Fine-tuning summary generation models with high-quality datasets lead to better models. The data filtering and segmentation are shown in Figure 2.
In Figure 2, the sentence filtering layer PureText performs data preprocessing on the original text (D) and the target text (T) to remove redundant sentences. Then, they are segmented to obtain text paragraphs ( D 1 , D 2 D n ) and ( T 1 , T 2 T n ) , respectively.

3.1.2. Coarse-Grained Summary Generation

During coarse-grained summary generation, the process of keyword extraction and integration is executed iteratively N times. Keywords are utilized to aid the inference process of summarization; a multistage approach is taken to extract coarse-grained level summary information from long texts. There are five main steps in the process, which are keyword extraction, self-attention accumulation calculation, cross-attention calculation, linear combination score calculation, and keyword incorporation into summary inference.
(1)
Keyword Extraction
The keyword extraction task is accomplished by using the keyword extraction module implemented in EmbedRank, which is capable of the lexical annotation of words. The words labeled NN, NNS, NNP, JJ, VB, VBD, etc., are recognized and the keywords are obtained by phrase extraction using Python’s own NLTK package. For example, for a given sentence ‘The process of creating experiments is minimized by the simple control of running the program code, achieved by editing most of the program’s parameters into a text file’, the keywords extracted from it are ‘program’s parameters’, ‘text file’, ‘program code’, ‘running’, ‘control’, ‘minimized’, and ‘process’. For the keywords initially extracted from the text, it is also necessary to use attention to calculate the importance of the keywords themselves as well as their relevance scores to the context of the sentence and the document, in order to obtain keywords that are of practical value for summary generation.
(2)
Self-Attention Accumulation Calculation
In the process of calculating self-attention accumulation for keywords, the input text segments ( D 1 , D 2 D n ) are lexically annotated and phrase marked to obtain a keyword set, K . The pre-trained BERT is utilized to extract w i k , which denotes the word self-attention weight of the keyword, k , in the i -th sentence. The accumulated self-attention weight, a i k , for the keyword, k , in the sentence, i , is then calculated. Finally, the self-attention of the keywords in each sentence is summed to obtain the document-level self-attention value, a k , for the keyword, k. The self-attention accumulation calculation is shown in Figure 3.
In Figure 3, a w is obtained from the pre-trained BERT, and a w denotes the value of the self-attention weights of the words in the sentence; the equation is shown below:
a w = w s a w w
where a w w denotes the value of the attentional weight of word w together with all the other words, w , in the same sentence, s .
Assuming that the sum of the self-attention weight values of keyword k in sentence i is a i k , the self-attention weight values of the words in k are summed to obtain a i k ; the equation is shown below:
a i k = w k a w
The document-level self-attention weight value, a k , of keyword k is the sum of the keyword weight values in all sentences; the equation is shown below:
a k = i d a i k
where d denotes the document and sentence i denotes the i -th sentence in the document.
The importance of the keywords is taken into account by calculating the self-attention accumulation of the keywords. However, the evaluation of the keywords is multifaceted and needs to be considered regarding the relevance of the keywords to the context of the sentence and the document.
(3)
Cross-Attention Calculation
To calculate the relevance scores of keywords to sentence context and documents, the network architecture based on the Hierarchical Attention Retrieval (HAR) model and the Bidirectional Attention model is used to design the cross-attention structure. It calculates the relevance between keywords and documents based on the context. A mixture of self-attention and cross-attention is used. The cross-attention calculation is shown in Figure 4.
In Figure 4, keyword k containing m words is denoted as E k = { e 1 k , , e m k } , with e i H denoting the embedding of w i . Similarly, sentence i containing n words is denoted as E i = { e 1 i , , e n i } . Calculating document embeddings through cross-attention provides a better measure of the contextual relevance between keywords and sentences in a document.
The equation of the similarity matrix, S n × m , between sentence i and keyword k is shown below:
S = E i E k T
where E i n × H denotes sentence i and E k m × H denotes keyword k .
The word-based sentence-to-keyword S ¯ i 2 k and keyword-to-sentence S ¯ k 2 i similarity calculations are shown in Equations (5) and (6), respectively:
S ¯ i 2 k = s o f t m a x r o w ( S )
S ¯ k 2 i = s o f t m a x c o l u m n ( S )
The calculation of cross-attention weights from sentence to keyword, A i 2 k , and word-based keyword to sentence, A k 2 i , are shown in Equations (7) and (8), respectively:
A i 2 k = S ¯ i 2 k E k
A k 2 i = S ¯ i 2 k S ¯ k 2 i T E i
Based on the cross-attention weights, the new sentence representation, V i , for sentence i is obtained by averaging; the equation is shown below:
V i = A V G ( E i , A i 2 k , E i A i 2 k , E i A k 2 i )
where E i denotes the original context of the sentence, A i 2 k , E i A i 2 k , and E i A k 2 i measure the contextual relevance between the sentence and the keywords, and denotes element-by-element multiplication. The new sentence representation is obtained by calculating the average of the sum of the four variables mentioned above, incorporating word-based relationships between keywords and sentences.
A new sentence representation containing n words is V i = { v 1 i , , v n i } , and to highlight the importance of the words, V i performs self-attention. The average of all columns is calculated to obtain the final representation of the sentence α i H . The self-attention and α i of sentence i are shown in Equations (10) and (11), respectively:
I = s o f t m a x r o w ( V i V i T ) V i
α i = A V E ( I [ : , i ] )
A document, d , containing a set of sentences is denoted as E d = { α 1 , , α i } , and E d is subjected to performing self-attention. The average of all columns is calculated as the final document embedding p d H , and the self-attention and p d of document d are shown in Equations (12) and (13), respectively:
P = s o f t m a x r o w ( E d E d T ) E d
p d = A V E ( P [ : , i ] )
The set of keyword embeddings is denoted as E k = { e 1 k , , e m k } , self-attention is applied to E k , and the average of all the columns is calculated to obtain the final keyword embedding, p k H . The self-attention and p k for keyword k are shown in Equations (14) and (15), respectively:
K = s o f t m a x r o w ( E k E k T ) E k
p k = A V E ( K [ : , i ] )
The equation of the relevance value, r k , between keyword k and document d is shown below:
r k = p k p d p k p d
where r k is determined by a cosine similarity calculation by p k and p d . It takes into account both the interrelationship between the keywords and the context of the sentence and the relevance between the keywords and the document, weighing the importance of the keywords from multiple perspectives.
(4)
Linear Combination Score Calculation
The importance of the keywords and the semantic relevance between the keywords and the document sentences are fully taken into account after self-attention accumulation calculation and cross-attention calculation for the preliminary extracted keywords. The final score, s k , of the keywords is calculated by the linear combination method; the equation is shown below:
s k = d a k + ( 1 d ) r k
where d [ 0 , 1 ] , the normalized self-attention value a k , and the cross-attention based relevance value r k are calculated accumulatively for each keyword, respectively. The final score of the keywords, s k , is determined by linearly combining a k and r k . The keywords are further evaluated based on the scores, and a portion of the useless keywords are eliminated. It more effectively integrates the keywords into the summary inference, resulting in an improvement in the quality of the summary.
(5)
Keyword Incorporation into Summary Inference
In the N-coarse-grained summary generation stages, redundant text information is removed through sentence filtering and data segmentation to further streamline the text content. The obtained data samples ( D i , T i ) are provided to the summary generation model to obtain a coarse-grained paragraph summary, Z i l , where l [ 1 , N ] , l denotes the index of the current stage; the equation is shown below:
Z i l = MSMAUKE - S UMM l ( D i , T i )
Concatenate n coarse-grained paragraph summaries ( Z 1 l , Z 2 l Z n l ) to obtain Z l = Z 1 l Z 2 l Z n l and use Z l as a new data sample for the next stage.
MSMAUKE-SummN can be applied to summary generation models such as BART and T5 [31], because the BART model performs poorly on summary generation for long texts. Therefore, in order to reflect the framework’s facilitation of BART for long-text summarization, BART is adopted as the backbone model of MSMAUKE-SummN.

3.2. 1-Fine-Grained Summary Generation Stage

When Z l is used as an input source, if the text length of Z l is less than the maximum input length, L , of the summary generation model, it can satisfy the requirements of the summary generation model in the 1-fine-grained stage. In the 1-fine-grained summary generation stage, the summary generation model is trained directly on the dataset ( Z N , T ) from the last coarse-grained stage.
Keyword Selection: a pre-trained encoder model with x self-attentive heads h 1 , h 2 , , h x is used in the 1-fine-grained summary generation stage. Each self-attentive head, hi, attends to a different part of the input text [32]. Divide each head into y clusters, C = { c 1 , c 2 , , c y } , so there are g = x y heads per cluster. The headers are clustered using a sequential approach and the m most popular keywords are identified from each header. Since a keyword may receive higher attention from more than one head, it may result in an overlap in the set of keywords obtained by each head. To solve the problem of overlapping keywords, up to g m keywords should be obtained from each cluster. Assume that the total number of keywords used for the coarse-grained summary inference process is n . By retaining r keywords as topic keywords, the remaining n r keywords are non-overlapping keywords with less attention. r is a hyperparameter, and it can be derived from the statistical laws that r is typically 10% of n . Similar to the inference process of using keywords for coarse-grained summarization, it is also necessary to further select a portion of keywords that are more relevant to the topic for use in the inference process of fine-grained summarization.
Keywords are utilized to assist the inference process of the summary generation model. A fine-grained summary, F , is obtained as the final output of MSMAUKE-SummN; the equation is shown below:
F = MSMAUKE - S UMM N + 1 ( Z N , T )
where ( Z N , T ) denotes the dataset of the last coarse-grained stage, and the dataset ( Z N , T ) is utilized by the MSMAUKE-SummN framework to finetune the summary generation model in the (N + 1)-th stage to obtain the final fine-grained summary, F .
The pseudo-code for MSMAUKE-SummN is shown in Algorithm 1:
Algorithm 1 MSMAUKE-SummN pseudo-code
Input: 
D , T ,   L r , Coarse-grained beam width, Fine-grained beam width, Input_max_token.
Output: 
F
1: for n in range(1, N + 1) do
2:     PureText ( D , T )
       Summary Generation Model   ROUGE ( D i , T i )
3:     keywords     NLTK ( D )
4:     a k     Accumulation ( keywords ) Self - attention
5:     r k     Relevance ( keywords ,   sentence ,   document ) Mixed - attention
6:     s k     Linear _ combination _ score ( a k , r k )
7:     Z i l     M S M A U K E - S U M M l ( D i , T i )
         ( Z 1 , Z 2 , Z n )   =   Z i l
8:     Z     Tan dem ( Z 1 , Z 2 , Z n )
       Summary Generation Model   ROUGE ( Z i , T i )
9: end for
10: F     M S M A U K E - S U M M N + 1 ( Z N , T )
11: return  F
Streamlined data samples are obtained by the sentence filtering process. Mixed-attention keyword extraction is utilized to enrich the global semantic information of coarse-grained summaries. The self-attentive keyword selection module is designed and keywords are selected to assist summary inference to enhance the local semantic representation of fine-grained summaries. An accurate, comprehensive, and readable fine-grained summary, F , is obtained.

4. Experimental Results and Analysis

4.1. Experimental Setup and Dataset

In this study, the experimental environment is configured on a cloud server with a single NVIDIA RTX A5000 (NVIDIA, Santa Clara, CA, USA) with 24 GB of video memory. The Python version 3.7 programming language, the Pytorch version 1.8.1 framework, CUDA version 11.1, and fairseq version 0.10.0 are utilized for the experiments.
The experimental hyperparameter settings are shown in Table 1.
From Table 1, Lr denotes the learning rate, which is set to 2 × 10−5. The coarse-grained beam width and fine-grained beam width denote the beam widths of the coarse-grained and fine-grained stages, which are set to 2 and 10, respectively. Input_max_token denotes the maximum number of tokens in the original text, which is set to 1024.
In addition, by looking at the model parameters, we can find that the BART-large model pre-trained on the CNN/DM dataset has the best summary generation capability [33]. Therefore, in this paper, the model parameters of BART-large trained on the CNN dataset are used as initialization parameters.
Both AMI and ICSI are conference scripts generated through automatic speech recognition systems. AMI is a dataset collected from a company’s product design meetings, and ICSI is a dataset collected from an academic group’s meetings. QMSum [34] is a query-based dataset of conference summaries, where each query and example are written by an expert. SummScreen [35] is a dataset consisting of transcripts of TV episodes from TVMegaSite (TMS) and ForeverDream (FD). The summary of each record is a summary of the TMS or FD. GovReport [36] is a large-scale long-document summary dataset containing 19,466 long reports issued by the U.S. Government Accountability Office on national policy issues. These datasets were chosen, in general terms, because they all belong to the long text summarization type of datasets. In terms of separate domains, they are from company meetings, academic conferences, long document reports released by the government, etc., which can improve the generalization ability of the MSMAUKE-SummN model and help the proposed summary generation model more. Therefore, these datasets are selected for related experiments in this study.
The statistics of the dataset used for the experiment are shown in Table 2.
The original text length and target text length are averaged over the entire dataset, and N denotes the number of coarse-grained summary generation stages.

4.2. Evaluation Metrics

ROUGE is used as an evaluation metric in this study, which is mainly categorized into ROUGE-1 (hereinafter referred to as R-1), ROUGE-2 (hereinafter referred to as R-2), and ROUGE-L (hereinafter referred to as R-L). R-1 is usually used as the standard evaluation metric because higher scores on the R-1 metric usually imply higher scores on the R-2 and R-L metrics as well. Meanwhile, R-1 has a lower time complexity compared to other ROUGE metrics.
The equation of definitions for R-1, R-2, and R-N are shown as follows:
R - N = R { R e f e r e n c e S u m m e r i e s } g r a m n R C o u n t m a t c h ( g r a m n ) R { R e f e r e n c e S u m m e r i e s } g r a m n R C o u n t ( g r a m n )
where n represents the length of the g r a m n , C o u n t m a t c h ( g r a m n ) denotes the maximum number of g r a m n that appear in both the candidate and reference summaries, and R denotes the reference summary.
For R-L, L is the longest common subsequence (LCS), and the LCS is used by R-L. R-L is calculated from R l c s , P l c s , and F l c s . R l c s and P l c s denote recall and precision, respectively, while F l c s denotes R-L. R-L is defined by Equations (21)–(23) as follows:
R l c s = L C S ( X , Y ) m
P l c s = L C S ( X , Y ) n
F l c s = ( 1 + β 2 ) R l c s P l c s R l c s + β 2 P l c s
where L C S ( X , Y ) is the length of the longest common subsequence of X and Y . m and n denote the lengths of the reference and auto-summaries, respectively. Since β is a very large number, R-L considers R l c s almost exclusively.

4.3. Ablation Experiments

To better validate the usefulness of the modules, the following ablation experiments are designed on the relatively long dataset GovReport. Only the sentence filtering layer PureText is added in Experiment 1. Only the mixed-attention keyword extraction module is incorporated to aid in summary inference in Experiment 2. The sentence filtering layer PureText and the mixed-attention keyword extraction module are added in Experiment 3. The mixed-attention keyword extraction module and the self-attention keyword selection module are incorporated in Experiment 4. The sentence filtering layer PureText, the mixed-attention keyword extraction module, and the self-attention keyword selection module are added in Experiment 5. The ablation experiment module selection is shown in Table 3.
The results of ablation Experiments 1, 2, 3, 4, and 5 on the three metrics R-1, R-2, and R-L are shown in Table 4.
As can be seen from Table 4, the addition of the sentence filtering layer PureText and the mixed-attention keyword extraction module in Experiments 1 and 2, respectively, resulted in an improvement in the R-1, R-2, and R-L metrics compared to SummN. The improvement was more obvious when the latter was added alone, with an improvement of 0.46%, 0.92%, and 0.81%, respectively. Sentence filtering provides a streamlined sample of data without any effective processing of keywords. Therefore, the improvement in summary generation metrics is relatively small, while the mixed-attention mechanism makes self-attention and cross-attention interact with each other, and fully focus on the keywords at the sentence and document levels. It is more beneficial for obtaining the important information of the summary and improving the readability and conciseness of the summary, and the improvement effect is more obvious.
In addition, from Experiment 3, it can be seen that with the simultaneous inclusion of the sentence filtering layer PureText and the mixed-attention keyword extraction module, the quality of summary generation can be improved by them in a mutually reinforcing way. From Experiment 4, it can be seen that the three metrics are improved by 0.68%, 1.27%, and 1.12% with the addition of the mixed-attention keyword extraction module and the self-attention keyword selection module, respectively. This is due to the large number of keywords initially extracted and the fact that the keywords were not filtered by methods such as attention, and there are many duplicates, which is not beneficial to the summarization inference process. Thus, it is not directly useful for summarization inference, while the mixed-attention weighting of keywords in a document is calculated implicitly, taking repetition into account. Semantic relevance between keywords and sentences in the document is also considered to enrich the global semantic information of the summary. On the other hand, keywords with higher weights can be found and utilized by the keyword selection module to assist summary inference. The ability of the summary generation model to represent detailed information is strengthened, and the local semantic representation of the summary is enhanced.
From Experiment 5, it is clear that the enhancement of the ROUGE metric by the three metrics is accumulative. Because redundant information in data samples can be reduced by sentence filtering, the redundancy of summary information can be reduced and the summary generation model can be improved. The summary generation model can be more comprehensively assisted by attention to obtain important keywords. The accuracy, comprehensiveness, and readability of summary information can be improved. Therefore, attention mechanisms can lead to greater enhancement.

4.4. Comparison Experiments

On the AMI, ICSI, and QMSum datasets, the improved modeling framework MSMAUKE-SummN is compared with the following comparison models in this paper:
  • HMNet [15]: A hierarchical attention structure and cross-domain pretraining are used by HMNet to extract meeting summaries, and a novel abstract summarization network is constructed to adapt to meeting scenarios. Role vectors are designed to cope with the semantic structure and style of different meeting records, and long meeting records are accommodated through a hierarchical structure.
  • TextRank [37]: A graph ranking approach is utilized by TextRank to consider the importance of each node’s information. It recursively calculates the importance weight value of each node in the relationship graph from the global semantic information. It counts and ranks important node information for keyword extraction and text summarization.
  • HAT-BART [16]: Layered attention is applied by HAT-BART to Transformer. Visual Hierarchical Encoder–Decoder Attention is designed to overcome the uneven distribution of important information in long texts. Hierarchical Attention Transformer’s (HAT) architecture is used to adapt to longer text inputs.
  • DDAMS [38]: Relational maps are used by DDAMS to model different discourse relationships; the interactions between discourses in a meeting are explicitly modeled. A relational graph encoder module is constructed, and graph interaction is used to model the interactions between discourses. The semantic relationships and information structure in the summaries are clarified.
  • SummN [17]: The data samples are segmented by SummN. It generates coarse-grained summaries through multiple stages, based on which fine-grained summaries are generated. The framework can handle input text of any length by adjusting the number of stages.
The results of the ROUGE metric values of the MSMAUKE-SummN model with HMNet, TextRank, HAT-BART, DDAMS, and SummN on the AMI, ICSI, and QMSum datasets are shown in Table 5.
For some special datasets, such as the SummScreen dataset, in this paper, the neural model Longformer + ATT and the hybrid model NN + BM25 + Neural proposed by Chen et al. [35] are used as comparison models and experimented on the SummScreen dataset. In addition, for the GovReport dataset, the comparison models are divided into two main categories, one of which is the BART variant, which represents the self-attentive variant with full attention. Another category is BART HEPOS, denoting the encoder variant with head position step encoder–decoder attention. The results of the ROUGE metric values of the MSMAUKE-SummN and other models on the SummScreen and GovReport datasets are shown in Table 6 and Table 7, respectively.
As can be seen in Figure 5, the model is significantly improved after adding the sentence filtering layer, the mixed-attention keyword extraction module, and the self-attention keyword selection module. The improvement effect is more obvious in Figure 5b compared to (a) because the ICSI dataset is relatively large and more information can be provided for summary generation. Meanwhile, MSMAUKE-SummN outperforms models such as SummN in all categories of ROUGE because sentence filtering long texts results in high-quality data samples, which helps to improve the quality of summary generation. The importance of keywords at the sentence level and document level can be paid full attention by the mixed-attention keyword extraction module. The extracted keywords are used in the stage of coarse-grained summary generation. The inference process of summary generation is enhanced, and the information extraction ability in the coarse-grained stage is better improved. In addition, keywords with higher weights are used by the self-attentive keyword selection module in the inference process of fine-grained summary generation. It is beneficial to further improve the summary ability of the summary generation model on detailed and focused information.
As can be seen in Figure 6, since SummScreen is a larger-volume dataset in the TV series genre, the MSMAUKE-SummN model was able to improve on each of the ROUGE metrics over the other comparison models. However, parameter optimization becomes more difficult due to computational resource constraints, resulting in it being a little harder for the model to improve metrics on larger datasets. It can still be well illustrated that MSMAUKE-SummN with BART as the backbone model has a better summary generation ability.
As can be seen in Figure 7, on the document dataset GovReport, the semantic relevance between keywords and the context of sentences and documents is fully considered due to the mixed-attention mechanism. The keywords are considered from multiple perspectives, and the importance of the keywords themselves as well as their semantic relevance to the documents are combined to assist summary inference. The MSMAUKE-SummN model has a better summary generation ability compared to other comparison models when dealing with document-like datasets.
In order to further demonstrate the superior performance of the MSMAUKE-SummN model relative to other models at different text lengths, text of different lengths is intercepted by us to test the performance of the model for summary generation. The performance difference between the models can be clearly seen through the summary generation time. The performance test of the model with different text lengths is shown in the following figure.
As can be seen from Figure 8, the unit of time on the vertical axis is seconds; text paragraphs of different lengths are intercepted for the document dataset GovReport, with text paragraphs of lengths such as 500, 1000, and 1500 words; and the performance test of the model’s summary generation is accomplished on these text paragraphs of different lengths. As can be seen from the figure, the time taken by each model to generate a summary is very close for shorter texts of 500 words; however, as the text length grows, it can be seen that the difference in performance between the models gradually widens. This is because the longer the length of the text, the more it restricts the model’s ability to generate summaries, unlike the MSMAUKE-SummN model because of its ability to flexibly adjust the number of stages to accommodate different lengths of text and to effectively take keywords into account. Therefore, the MSMAUKE-SummN model can outperform other models in terms of performance and the quality of summary generation.
At different text lengths, there are large differences in both the time to process the text and the quality of the summaries generated by the different models. When the same length of text is processed, the MSMAUKE-SummN model will take less time compared to HAT-BART, DDAMS, and other models, and the difference between the MSMAUKE-SummN model and the other models will be more obvious as the length of the text grows. Thus, this is good proof of the advantages of the MSMAUKE-SummN model in processing long text summaries. Meanwhile, the MSMAUKE-SummN model proposed in this paper is able to adjust the number of generation stages to adapt to different lengths of text, whereas models such as HAT-BART have a more complex and hierarchical structure, which is less adaptive to the task of summarizing long texts. Therefore, the MSMAUKE-SummN model is characterized by a simple and flexible structure compared to other models. Similarly, the MSMAUKE-SummN model in this paper is tediously trained on various types of long-text summarization datasets, which sufficiently improves the model’s adaptability to different domains and provides a better generalization of the MSMAUKE-SummN model compared to other models that are trained on more homogeneous datasets.

5. Conclusions

In the text generation task, the generation of summaries for long texts is an important task; an N + 1 coarse–fine-grained multistage summary generation framework is constructed and a multistage mixed-attention unsupervised keyword extraction summary generation model (MSMAUKE-SummN) is proposed. In the N-coarse-grained summary generation stage, the sentence filtering layer, PureText, is constructed to remove redundant information from the text. It improves the summary generation model’s ability to extract important information. Self-attention and cross-attention are mixed. The importance of keywords at the sentence level and document level is fully considered, and keywords are extracted stage-by-stage in an unsupervised manner using mixed attention. The inference process of coarse-grained summary generation is aided and the global semantic information of coarse-grained summaries is enriched. In the 1-fine-grained summary generation stage, the self-attentive keyword selection module is designed to improve the controllability of keyword usage. Keywords with higher weights are incorporated into the summary inference. The local semantic representation of the fine-grained summary is enhanced. By flexibly adapting to texts of different lengths through the N + 1 coarse–fine-grained multistage summary generation framework, the quality of summaries is considered from both global and local semantic perspectives. The accuracy, comprehensiveness, and readability of summary information are improved, and the redundancy of summary information is reduced. The experimental results show that the MSMAUKE-SummN model framework is improved over HMNET, TextRank, HAT-BART, DDAMS, and SummN models in terms of ROUGE-1, ROUGE-2, and ROUGE-L metrics, with a minimum of 0.75%, 1.48%, and 1.25% improvement in the three metrics across the datasets relative to the SummN model. It is demonstrated that MSMAUKE-SummN has a better summary generation ability for long texts or documents.
Further research will be conducted in future work on how to generate summaries of multimodal data, which can be data that contains multiple forms such as text and audio, text and image, or text and video, as opposed to a single text type of data. The fusion of MSMAUKE-SummN with various improved convolutional neural networks is considered in future work to investigate how to improve applicability to multi-domain data. In addition, how to further improve the accuracy of keyword extraction is also to be solved. In subsequent research work, diversified feature weighting will be considered to improve the self-attention mechanism, and multiple features of keywords are considered comprehensively to improve the accuracy of keyword extraction.

Author Contributions

D.W., P.C. and Y.Z.: conceptualization; P.C.: validation; D.W., P.C. and Y.Z. wrote the paper; D.W. and Y.Z. reviewed the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Research Projects of the Nature Science Foundation of Hebei Province (No. F2021402005), as well as the National Natural Science Foundation of China (No. 62101174).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to privacy.

Acknowledgments

The authors look forward to the insightful comments and suggestions of the anonymous reviewers and editors, which will go a long way towards improving the quality of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, H.; Liu, X.; Zhang, J. Extractive Summarization via ChatGPT for Faithful Summary Generation. In Findings of the Association for Computational Linguistics: EMNLP; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 3270–3278. [Google Scholar]
  2. Bao, G.; Ou, Z.; Zhang, Y. GEMINI: Controlling The Sentence-Level Summary Style in Abstractive Text Summarization. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Singapore, 6–10 December 2023; pp. 831–842. [Google Scholar]
  3. Chen, J.; Yang, D. Structure-Aware Abstractive Conversation Summarization via Discourse and Action Graphs. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 1380–1391. [Google Scholar]
  4. Yoo, C.; Lee, H. Improving Abstractive Dialogue Summarization Using Keyword Extraction. Appl. Sci. 2023, 13, 9771. [Google Scholar] [CrossRef]
  5. Chen, T.; Wang, X.; Yue, T. Enhancing Abstractive Summarization with Extracted Knowledge Graphs and Multi-Source Transformers. Appl. Sci. 2023, 13, 7753. [Google Scholar] [CrossRef]
  6. Xia, W.; Huang, H.; Gengzang, C.; Fan, Y. A review of extractive text summarisation based on unsupervised and supervised learning. Comput. Appl. 2023, 1–17. [Google Scholar]
  7. Zhang, L.; Chen, Q.; Wang, W.; Deng, C.; Zhang, S.; Li, B.; Wang, W.; Cao, X. MDERank: A Masked Document Embedding Rank Approach for Unsupervised Keyphrase Extraction. In Findings of the Association for Computational Linguistics: ACL; Association for Computational Linguistics: Stroudsburg, PA, USA, 2022; pp. 396–409. [Google Scholar]
  8. Sun, Y.; Qiu, H.; Zheng, Y.; Wang, Z.; Zhang, C. SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-Trained Language Model. IEEE Access 2020, 8, 10896–10906. [Google Scholar] [CrossRef]
  9. Saxena, A.; Mangal, M.; Jain, G. KeyGames: A Game Theoretic Approach to Automatic Keyphrase Extraction. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 2037–2048. [Google Scholar]
  10. Leonardo, F.R.R.; Mohit, B.; Markus, D. Generating Summaries with Controllable Readability Levels. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Singapore, 6–10 December 2023; pp. 11669–11687. [Google Scholar]
  11. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  12. Griffin, A.; Alex, F.; Faisal, L.; Eric, L.; Noémie, E. From Sparse to Dense: GPT-4 Summarization with Chain of Density Prompting. In Proceedings of the 4th New Frontiers in Summarization Workshop, Singapore, 6–10 December 2023; pp. 68–74. [Google Scholar]
  13. Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. BART: Denoising Sequence-to-Sequence Pre-Training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7871–7880. [Google Scholar]
  14. Kitaev, N.; Kaiser, L.; Levskaya, A. Reformer: The Efficient Transformer. International Conference on Learning Representations. arXiv 2020, arXiv:2001.04451. [Google Scholar]
  15. Zhu, C.; Xu, R.; Zeng, M.; Huang, X. A Hierarchical Network for Abstractive Meeting Summarization with Cross-Domain Pretraining. In Findings of the Association for Computational Linguistics: EMNLP; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 194–203. [Google Scholar]
  16. Rohde, T.; Wu, X.; Liu, Y. Hierarchical Learning for Generation with Long Source Sequences. arXiv 2021, arXiv:2104.07545. [Google Scholar]
  17. Zhang, Y.; Ni, A.; Mao, Z.; Wu, C.H.; Zhu, C.; Deb, B.; Awadallah, A.; Radev, D.; Zhang, R. SummN: A Multi-Stage Summarization Framework for Long Input Dialogues and Documents. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Volume 1, pp. 1592–1604. [Google Scholar]
  18. Fang, J.; Li, B.; You, X.; Lv, X. PLSGA: A staged approach to long text summary generation. Comput. Eng. Appl. 2023, 1–10. [Google Scholar]
  19. Ren, S.; Zhang, J.; Zhap, Z.; Rao, D. A two-stage text summarization model combining topic and location information. Intell. Comput. Appl. 2023, 13, 158–163. [Google Scholar]
  20. Mei, A.; Kabir, A.; Bapat, R.; Judge, J.; Sun, T.; Wang, W.Y. Learning to Prioritize: Precision-Driven Sentence Filtering for Long Text Summarization. In Proceedings of the 13th Language Resources and Evaluation Conference, Marseille, France, 20–25 June 2022; pp. 313–318. [Google Scholar]
  21. Sun, S.; Liu, Z.; Xiong, C.; Liu, Z.; Bao, J. Capturing Global Informativeness in Open Domain Keyphrase Extraction. In Proceedings of the Natural Language Processing and Chinese Computing: 10th CCF International Conference, NLPCC 2021, Qingdao, China, 13–17 October 2021. [Google Scholar]
  22. Song, M.; Jing, L.; Xiao, L. Importance Estimation from Multiple Perspectives for Keyphrase Extraction. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021; pp. 2726–2736. [Google Scholar]
  23. Papagiannopoulou, E.; Tsoumakas, G.; Papadopoulos, A. Keyword Extraction Using Unsupervised Learning on the Document’s Adjacency Matrix. In Proceedings of the Fifteenth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-15), Mexico City, Mexico, 11 June 2021; pp. 94–105. [Google Scholar]
  24. Kong, A.; Zhao, S.; Chen, H.; Li, Q.; Qin, Y.; Sun, R.; Bai, X. PromptRank: Unsupervised Keyphrase Extraction Using Prompt. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, Toronto, ON, Canada, 10–12 July 2023; Volume 1, pp. 9788–9801. [Google Scholar]
  25. Joshi, R.; Balachandran, V.; Saldanha, E.; Glenski, M.; Volkova, S.; Tsvetkov, Y. Unsupervised Keyphrase Extraction via Interpretable Neural Networks. In Findings of the Association for Computational Linguistics: EACL; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 1107–1119. [Google Scholar]
  26. Ding, H.; Luo, X. AttentionRank: Unsupervised Keyphrase Extraction using Self and Cross Attentions. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021; pp. 1919–1928. [Google Scholar]
  27. Kreutz, T.; Daelemans, W. Streaming Language-Specific Twitter Data with Optimal Keywords. In Proceedings of the 12th Web as Corpus Workshop, Marseille, France, 11–16 May 2020; European Language Resources Association: Luxemburg, 2020; pp. 57–64. [Google Scholar]
  28. Akash, P.S.; Huang, J.; Chang, K.; Li, Y.; Popa, L.; Zhai, C. Domain Representative Keywords Selection: A Probabilistic Approach. In Findings of the Association for Computational Linguistics: ACL; Association for Computational Linguistics: Stroudsburg, PA, USA, 2022; pp. 679–692. [Google Scholar]
  29. Venkatesh, E.; Kaushal, M.; Deepak, K.; Maunendra, S.D. DivHSK: Diverse Headline Generation using Self-Attention based Keyword Selection. In Findings of the Association for Computational Linguistics: ACL; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 1879–1891. [Google Scholar]
  30. Sreyan, G.; Utkarsh, T.; Manan, S.; Sonal, K.; Ramaneswaran, S.; Dinesh, M. ACLM: A Selective-Denoising based Generative Data Augmentation Approach for Low-Resource Complex NER. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, Toronto, ON, Canada, 9–14 July 2023; Volume 1, pp. 104–125. [Google Scholar]
  31. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  32. Peng, H.; Schwartz, R.; Li, D.; Smith, N.A. A Mixture of h − 1 Heads is Better than h Heads. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 6566–6577. [Google Scholar]
  33. Zhang, Y.; Ni, A.; Yu, T.; Zhang, R.; Zhu, C.; Deb, B.; Celikyilmaz, A.; Awadallah, A.H.; Radev, D. An Exploratory Study on Long Dialogue Summarization: What Works and What’s Next. arXiv 2021, arXiv:2109.04609. [Google Scholar]
  34. Zhong, M.; Yin, D.; Yu, T.; Zaidi, A.; Mutuma, M.; Jha, R.; Awadallah, A.H.; Celikyilmaz, A.; Liu, Y.; Qiu, X.; et al. QMSum: A New Benchmark for Query-based Multi-domain Meeting Summarization. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 5905–5921. [Google Scholar]
  35. Chen, M.; Chu, Z.; Wiseman, S.; Gimpel, K. SummScreen: A Dataset for Abstractive Screenplay Summarization. arXiv 2022, arXiv:2104.07091. [Google Scholar]
  36. Huang, L.; Cao, S.; Parulian, N.; Ji, H.; Wang, L. Efficient Attentions for Long Document Summarization. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 1419–1436. [Google Scholar]
  37. Mihalcea, R.; Tarau, P. TextRank: Bringing Order into Text. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; pp. 404–411. [Google Scholar]
  38. Feng, X.; Feng, X.; Qin, B.; Geng, X. Dialogue discourse-aware graph model and data augmentation for meeting summarization. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 19–27 August 2021; pp. 3808–3814. [Google Scholar]
Figure 1. The framework of the MSMAUKE-SummN model.
Figure 1. The framework of the MSMAUKE-SummN model.
Applsci 14 02435 g001
Figure 2. The data filtering and segmentation.
Figure 2. The data filtering and segmentation.
Applsci 14 02435 g002
Figure 3. The self-attention accumulation calculation.
Figure 3. The self-attention accumulation calculation.
Applsci 14 02435 g003
Figure 4. The cross-attention calculation.
Figure 4. The cross-attention calculation.
Applsci 14 02435 g004
Figure 5. ROUGE metric values on AMI (a) and ICSI (b) datasets.
Figure 5. ROUGE metric values on AMI (a) and ICSI (b) datasets.
Applsci 14 02435 g005
Figure 6. ROUGE metric values on SummScreen-FD (a) and SummScreen-TMS (b) datasets.
Figure 6. ROUGE metric values on SummScreen-FD (a) and SummScreen-TMS (b) datasets.
Applsci 14 02435 g006
Figure 7. ROUGE metric values on the GovReport dataset for comparative models such as LSH (4096) and Sinkhorn (5120).
Figure 7. ROUGE metric values on the GovReport dataset for comparative models such as LSH (4096) and Sinkhorn (5120).
Applsci 14 02435 g007
Figure 8. Performance test of the model with different text lengths.
Figure 8. Performance test of the model with different text lengths.
Applsci 14 02435 g008
Table 1. Experimental hyperparameter settings.
Table 1. Experimental hyperparameter settings.
Serial NumberParametersParameter Value
1Lr2 × 10−5
2Coarse-grained beam width2
3Fine-grained beam width10
4Input_max_token1024
Table 2. Statistics of the dataset.
Table 2. Statistics of the dataset.
DatasetTypeDomainSizeOriginal Text LengthTarget Text LengthN + 1
AMIDialogueMeetings1376007.7296.62
ICSIDialogueMeetings5913,317.3488.53
QMSumDialogueMeetings18089069.869.62
SummScreenDialogueTV shows26,8516612.5337.42
GovReportDocumentReports19,4669409.4553.43
Table 3. Selection of experimental modules for MSMAUKE-SummN ablation experiments.
Table 3. Selection of experimental modules for MSMAUKE-SummN ablation experiments.
Experiment NumberSentence
Filtering
Mixed-Attention Keyword
Extraction
Self-Attention Keyword
Selection
1
2
3
4
5
A “√” in the above table indicates that the corresponding module has been selected.
Table 4. Results of MSMAUKE-SummN ablation experiments.
Table 4. Results of MSMAUKE-SummN ablation experiments.
Evaluation MetricsR-1 (%)R-2 (%)R-L (%)
Experiment Number
156.8723.5654.12
257.2324.1754.71
357.3524.3354.85
457.4524.5255.02
557.5224.7355.15
MSMAUKE-SummN’s ROUGE metric values on the dataset GovReport are bolded.
Table 5. ROUGE metric values on AMI, ICSI and QMSum datasets.
Table 5. ROUGE metric values on AMI, ICSI and QMSum datasets.
DatasetAMIICSIQMSum-AllQMSum-Gold
Model R-1
(%)
R-2
(%)
R-L
(%)
R-1
(%)
R-2
(%)
R-L
(%)
R-1
(%)
R-2
(%)
R-L
(%)
R-1
(%)
R-2
(%)
R-L
(%)
HMNET52.3618.6324.0045.9710.1418.5432.298.6728.1736.0611.3631.27
TextRank35.196.1316.7030.724.6912.9716.272.6915.41---
HAT-BART52.2720.1550.5743.9810.8341.36------
DDAMS53.1522.3225.6740.4111.0219.18------
SummN53.4420.3051.3945.5711.4943.3234.039.2829.4840.2015.3235.62
MSMAUKE-SummN55.1921.4653.3547.5412.8545.2536.2910.831.2242.217.0938.09
MSMAUKE-SummN’s ROUGE metric values on the AMI, ICSI, and QMSum datasets are bolded.
Table 6. ROUGE metric values on the SummScreen dataset.
Table 6. ROUGE metric values on the SummScreen dataset.
DatasetSummScreen-FDSummScreen-TMS
Model R-1 (%)R-2 (%)R-L (%)R-1 (%)R-2 (%)R-L (%)
Longformer + ATT25.94.223.842.911.941.6
NN + BM25 + Neural25.33.923.138.810.236.9
SummN32.485.8527.5544.6411.8742.53
MSMAUKE-SummN34.346.8628.9846.3812.6444.74
MSMAUKE-SummN’s ROUGE metric values on the SummScreen dataset are bolded.
Table 7. ROUGE metric values on the GovReport dataset.
Table 7. ROUGE metric values on the GovReport dataset.
DatasetGovReport
Model R-1 (%)R-2 (%)R-L (%)
BART Variants
Full (1024)52.8320.550.14
Stride (4096)54.2920.851.35
LIN. (3072)44.8413.8741.94
LSH (4096)54.7521.3651.27
Sinkhorn (5120)55.4521.4552.48
BART HEPOS
LSH (7168)5521.1351.67
Sinkhorn (10,240)56.8622.6253.82
SummN56.7723.2553.9
MSMAUKE-SummN57.5224.7355.15
MSMAUKE-SummN’s ROUGE metric values on the GovReport dataset are shown in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, D.; Cheng, P.; Zheng, Y. Multistage Mixed-Attention Unsupervised Keyword Extraction for Summary Generation. Appl. Sci. 2024, 14, 2435. https://doi.org/10.3390/app14062435

AMA Style

Wu D, Cheng P, Zheng Y. Multistage Mixed-Attention Unsupervised Keyword Extraction for Summary Generation. Applied Sciences. 2024; 14(6):2435. https://doi.org/10.3390/app14062435

Chicago/Turabian Style

Wu, Di, Peng Cheng, and Yuying Zheng. 2024. "Multistage Mixed-Attention Unsupervised Keyword Extraction for Summary Generation" Applied Sciences 14, no. 6: 2435. https://doi.org/10.3390/app14062435

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop