Next Article in Journal
Hybrid Control-Based Closed-Loop Soft Start-Up Method for LLC Resonant Converters
Previous Article in Journal
Analysis of High-Frequency Communication Channel Characteristics in a Typical Deep-Sea Incomplete Sound Channel
Previous Article in Special Issue
A Multi-Modal Retrieval Model for Mathematical Expressions Based on ConvNeXt and Hesitant Fuzzy Set
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RoBERTa-Based Keyword Extraction from Small Number of Korean Documents

1
School of Computing, Kyung Hee University, Yongin 17104, Republic of Korea
2
Korea Electric Power Research Institute, Daejeon 34056, Republic of Korea
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(22), 4560; https://doi.org/10.3390/electronics12224560
Submission received: 20 September 2023 / Revised: 5 November 2023 / Accepted: 6 November 2023 / Published: 7 November 2023
(This article belongs to the Special Issue Natural Language Processing and Information Retrieval)

Abstract

:
Keyword extraction is the task of identifying essential words in a lengthy document. This process is primarily executed through supervised keyword extraction. In instances where the dataset is limited in size, a classification-based approach is typically employed. Therefore, this paper introduces a novel keyword extractor based on a classification approach. The proposed keyword extractor comprises three key components: RoBERTa, a keyword estimator, and a decision rule. RoBERTa encodes an input document, the keyword estimator calculates the probability of each token in the document becoming a keyword, and the decision rule ultimately determines whether each token is a keyword based on these probabilities. However, training the proposed model with a small dataset presents two challenges. One problem is the case that all tokens in the documents are not a keyword, and the other problem is that a single word can be composed of keyword tokens and non-keyword tokens. Two novel heuristics are thus proposed to tackle these problems. To address these issues, two novel heuristics are proposed. These heuristics have been extensively tested through experiments, demonstrating that the proposed keyword extractor surpasses both the generation-based approach and the vanilla RoBERTa in environments with limited data. The efficacy of the heuristics is further validated through an ablation study. In summary, the proposed heuristics have proven to be effective in developing a supervised keyword extractor with a small dataset.

1. Introduction

Keyword extraction is a task of identifying keywords in a lengthy document, and it is considered one of the essential tasks in data analysis, data mining, and information retrieval [1,2,3,4]. In the current environment, where countless documents are published through various media [5,6], it becomes time-consuming for humans to process all documents and identify their keywords. Consequently, the demand for automatic keyword extraction has increased [7,8,9].
Automatic keyword extraction has been addressed using unsupervised and supervised methods [10]. Two representative examples of unsupervised keyword extraction are Tf-idf [11] and PageRank [12,13]. Tf-idf detects keywords based on word frequency [14,15,16], while PageRank utilizes graph-based features [17,18,19]. However, these methods, lacking an understanding of the overall semantic information of a document, generally achieve relatively low accuracy. To address this issue, supervised keyword extraction has been introduced, reporting high performance [20]. Nevertheless, these methods typically necessitate a large amount of training data, posing a challenge when only limited data are available.
Supervised keyword extraction has been studied through two approaches: the classification-based approach and the generation-based approach. The classification-based approach involves extracting keywords from a document by evaluating every token in the document to determine if it is a keyword or not [21,22,23]. In contrast, the generation-based approach uses a generative language model to abstractively generate keywords for an input document [24,25]. According to numerous studies [25], generative language models like BART [26] outperform classification-based methods in extraction accuracy, making the generation-based approach more commonly adopted for keyword extraction. However, a drawback of generative language models is their requirement of a large amount of training data, surpassing that needed by classification-based methods in general.
This paper proposes a novel keyword extractor designed to be trained with a small number of Korean documents. The extractor addresses three problems that are challenging for generative language models. Firstly, due to the unavailability of large-scale keyword-spotting documents for minor languages such as Korean, the extractor must be trained with a limited amount of data. Consequently, the proposed keyword extractor adopts the classification-based approach [27] instead of the generation-based approach, utilizing RoBERTa [28], a BERT-based classifier, as a base model due to its outstanding performance among classification-based methods.
The second problem arises from RoBERTa determining whether every token in a document is a keyword or not. Consequently, it is highly possible that RoBERTa may not produce any keywords for a document since it is a token-level classifier. To address this, a heuristic is employed to identify the word to which the token with the highest probability belongs as a keyword. Additionally, since every word is expressed as a sequence of tokens in RoBERTa, owing to its BPE tokenizer, even a single word can be composed of multiple tokens with varying keyword predictions. The third problem is to unify the keyword predictions of multiple tokens for a single word. To address this issue, a heuristic is applied, considering a word as a keyword only when all its tokens are predicted as keywords, leveraging RoBERTa’s reliability in keyword extraction.
Intensive experiments are conducted using a dataset of Korean Power Plant Outage Reports. Although the dataset is small, the experiments demonstrate that RoBERTa achieves high performance compared to BART, a generative language model. Moreover, it is evident that vanilla RoBERTa faces the second and third problems, which the proposed heuristics effectively solve. Particularly, the proposed keyword extractor achieves an average 6.4 and 12.8 higher BLEU and accuracy, respectively, compared to vanilla RoBERTa. Furthermore, it outperforms vanilla RoBERTa even in human evaluation, affirming the efficacy of the proposed keyword extractor for datasets with a limited number of training instances.
The main contributions of this work are
  • This paper analyzes problems that may occur when a supervised classification-based keyword extractor is trained with a small number of data.
  • This paper proposes two heuristics for sophisticated keyword extraction to solve problems of classification-based keyword extractors trained with a small number of data.
  • The proposed keyword extractor outperforms not only the generation-based keyword extractor but also the classification-based keyword extractor in the environment of a small number of data.
The rest of this paper is organized as follows. Section 2 introduces the previous studies on automatic keyword extraction. Section 3 describes two problems in learning a RoBERTa-based keyword extractor with a small number of documents and addresses how to tackle these problems, and Section 4 provides the experimental results. Finally, Section 5 indicates the conclusions and future work.

2. Related Work

Traditionally, keyword extraction has been studied in an unsupervised way. Tf-idf [11,29] is a representative unsupervised method, where it detects keywords statistically based on the frequency of the tokens in a document. The methods of extracting keywords using term frequency like tf-idf require a large corpus to improve their performance. Thus, efforts to use not only term frequency but also some other features have been undertaken in order to solve this problem. KP-Miner utilized word length and positional information in addition to term frequency [30], RAKE considered the frequency of other tokens that a term co-occurs with [31], and YAKE calculated a keyword score with the term frequency and some hand-crafted features [8].
Graph-based keyword extraction is another way to extract keywords unsupervisedly with purely statistical information. TextRank [12,32] is a representative graph-based approach, where it builds an undirected, unweighted word graph representation and then finds out the keywords using the PageRank algorithm. On the other hand, SGRank [33] uses some manually crafted features in addition to TextRank, where the features are designed to complement TextRank since some critical information for keyword extraction cannot be captured by a graph-based method. KeyphraseDS [34] filters out non-keywords from the keyword candidates before TextRank using the similarity of a candidate to a document topic to keep stopwords from becoming keywords. These unsupervised methods have an advantage of requiring small data and faster inference, but their accuracy is relatively low because they extract keywords without a semantic understanding of a document. For this reason, supervised-based methods that extract keywords grasping the semantics of a document have begun to attract the attention of the keyword extraction research community.
Supervised keyword extraction is further divided into a classification-based approach and a generation-based approach. The classification-based approach is also called a sequence labeling approach. That is, it aims to determine whether each input token in a document is a keyword or not. For instance, Luan et al. [35] determined keywords with part-of-speech embeddings and the combinatorial features of words and characters. On the other hand, Chun et al. [36] proposed a novel embedding for sequence labeling, which integrates character-level embedding and word-level embedding to find the most intuitive character or word in a document. However, such classification-based methods always have a possibility that all tokens in a document are determined to be non-keywords.
Many studies have attempted to solve the problems of the classification-based approach by adopting a generative model based on a sequence-to-sequence architecture. This is because the generation-based approach generates keywords abstractly from a document and thus it outputs some keywords anyway. In the early studies, RNN [37] was used as a dominant keyword generator. However, many attempts have been made to improve keyword extraction performance recently by adopting a generative language model. For example, BART [26], a transformer-based language model, showed a higher performance than RNN [25]. Recently, there have also been efforts to generate keywords using prompts [38]. However, the main drawback of this generation-based approach is that it requires a great volume of data for its training. As a result, this paper takes the classification-based approach to solve keyword extraction for a dataset with a small number of documents.

3. Keyword Extraction with Small Number of Documents

3.1. Supervised Keyword Extraction

Keyword extraction is a task to extract the most relevant words from a document W = [ w 1 , , w m ] , a sequence of word w i ’s. This task can be solved in various ways. One way is the generation-based approach in which the keywords for an input document W are generated by a generative language model. That is, a generative language model generates Y = [ y ^ 1 , , y ^ o ] , where o is the targeted number of keywords. One benefit of this approach is that it can generate even the words that do not appear in W as keywords. However, it has a disadvantage of requiring a large amount of data to train the generative language model.
The other way to keyword extraction is the classification-based approach. This approach is usually formulated as sequence labeling. That is, keyword extraction in this approach can be considered as a series of word-level classifications. Under this approach, the keywords for W are extracted by determining whether every w i W is a keyword or not. In order to train a word-level classifier, a sequence of golden labels Y = [ y 1 , , y m ] for each W should be given where | X | = | Y | . In addition, all y i ’s are boolean so that y i is keyword if the corresponding w i is a keyword and non-keyword otherwise. Formally, let f ( · ) be a word-level classifier trained with a dataset D = { ( W i , Y i ) } i = 1 N . Then, when a new document W ¯ = [ w ¯ 1 , , w ¯ m ] is given, its label Y ¯ = [ y ¯ 1 , , y ¯ m ] is determined by applying f ( · ) to every w ¯ i W ¯ . That is, y ¯ i is obtained by
y ¯ i = f ( w ¯ i ; W ¯ ) .
Here, W ¯ is used as a context for this classification.
This paper adopts a RoBERTa-based model as the classifier f ( · ) in Equation (1) since RoBERTa achieves good performances in many sequence labeling tasks [39,40]. The structure of the proposed classifier is depicted in Figure 1. It consists of three components of RoBERTa, a keyword estimator, and a decision rule. The input document W is first encoded as vectors by RoBERTa. Since RoBERTa is a BERT-based model, every word in W is further divided into a few tokens by the byte-level BPE, a standard RoBERTa tokenizer. Thus, the document W = [ w 1 , , w m ] is represented as a sequence of tokens X = [ x 1 , , x n ] where n > m . Then, the task of keyword extraction becomes predicting Y = [ y 1 , , y n ] from X . Following the standard RoBERTa application, the [cls] token is assumed to be in X and is regarded as x 0 .
For every x i X , a corresponding vector representation h i is obtained by
h i = R o B E R T a ( x i ; X ) R E ,
where E is the embedding size. Note that RoBERTa is encoder-centric since it is based on BERT. Thus, an additional keyword estimator is required to adopt RoBERTa in an end-to-end style [22,23]. This paper uses a linear projection with softmax as the keyword estimator c ( · ) where it takes h i as an input and calculates the probability of h i ’s being a keyword. That is,
c ( h i ) = softmax ( h i W + B ) ,
where W R E × C and B R n × C are learnable parameters and C is the number of class labels. Since y ¯ i in Equation (1) can have only two values of keyword and non-keyword, C is two in this paper. This implies that the number of nodes in the softmax of c ( · ) is also two. Thus, c ( h i ) in Equation (2) returns p i = p i T , p i F , a vector of two probabilities. p i T is the probability that x i becomes a keyword and p i F is the probability that x i is not a keyword. As a result, p i F = 1 p i T holds. In a word,
p i = c ( h i ) = c ( R o B E R T a ( x i ; X ) ) .
The final keyword prediction is then made from p i by a simple rule
R ( p i ) = keyword if p i T p i F , non - keyword otherwise .
Therefore, y i is determined by
y i = f ( x i ; X ) = R ( c ( R o B E R T a ( x i ; X ) ) ) .
The parameters of W and B of the keyword estimator c ( · ) in Equation (2) are optimized with the training dataset D. RoBERTa is also fine-tuned with D.

3.2. Sequence Labeling for Small Number of Documents

Note that the size of the training data D should be large to obtain a good performance with the proposed keyword extractor in Equation (5). However, minor languages such as Korean usually do not have a large dataset labeled for keyword extraction. If the size of D is very small, the keyword extractor can suffer from two kinds of problems. One problem is that the keyword extractor is underfitted to the training data D and shows a deteriorated performance. One possible phenomenon by the underfitted keyword extractor is that all words in a document are predicted to be non-keywords since there are usually many more non-keywords than keywords in a document. The other problem is that a single word can be predicted to be both a keyword and a non-keyword. This is because the keyword extractor is based on RoBERTa in which a word is tokenized into multiple tokens. As a result, there could be a case that some tokens of a single word are classified as a keyword while other tokens are determined as a non-keyword.
The first problem is solved by adopting a simple heuristic, which is to consider the token with the highest probability as a keyword if all tokens are predicted as non-keywords. That is, if all y i ’s are non-keywords, i.e., p i T < p i F for 1 i n , then y q is forced to be a keyword where q is chosen by
q = arg max i { 1 , , n } p i T .
As a result, at least one token survives as a keyword.
This paper proposes two post-processings of reinforcement and exclusion as a solution to the second problem. Assume that a word w i is composed of k tokens of t i 1 , , t i k and the keyword label for w i is provided as [ y i 1 , , y i k ] , where y i j = f ( t i j ; X ) . Some y i j ’s are keyword and the remaining are non-keyword since the classifications for t i j ’s are performed independently. The reinforcement post-processing enforces w i to be a keyword even when some y i j ’s in w i are non-keyword. Algorithm 1 and Figure 2a show how this post-processing works in detail. If there is at least one token of which label is keyword, the word w i is forced to be keyword. Thus, in lines 3 and 4, the algorithm checks if there is any token t i j satisfying f ( t i j ; X ) = keyword . On the other hand, the exclusion post-processing enforces w i to be a keyword only when all y i j ’s are keyword. Algorithm 2 and Figure 2b show the flow of the exclusion post-processing. Similar to Algorithm 1, this algorithm checks if there is any token t i j satisfying f ( t i j ; X ) =  non-keyword in lines 3 and 4. Thus, if there is at least one non-keyword token, w i also becomes non-keyword.
Algorithm 1 The algorithm of reinforcement post-processing
input: a word w i = [ t i 1 , , t i k ]
output: keyword or non-keyword
 1:
is_keyword  non-keyword
 2:
for  1 j k  do
 3:
    if  f ( t i j ; X ) =  keyword then
 4:
         is_keyword keyword
 5:
    end if
 6:
end for
 7:
return  is_keyword
Algorithm 2 The algorithm of exclusion post-processing
input: a word w i = [ t i 1 , , t i k ]
output: keyword or non-keyword
 1:
is_keyword keyword
 2:
for  1 j k  do
 3:
    if  f ( t i j ; X ) =  non-keyword then
 4:
         is_keyword  non-keyword
 5:
    end if
 6:
end for
 7:
return  is_keyword

4. Experiments

4.1. Dataset

To prove that the proposed keyword extractor solves the problems that easily occur in keyword extraction supervised with small number of documents, a corpus named as Korean Power Plant Outage Reports is adopted. Every report in the corpus has three sections of ‘Failure Detail’, ‘Cause’, and ‘Measurement’. One dataset is made from one section. As a result, three datasets of failure detail, cause, and measurement are finally prepared from the corpus. Table 1 shows an example of the data.
The keywords in a data instance are tagged from two points of view. One point of view is about failure object and the other is about failure phenomenon. Thus, every data instance has two types of keywords: object and phenomenon. In the dataset of failure detail, all instances have only one object keyword and one phenomenon keyword. On the other hand, in cause and measurement, the instances have three or fewer keywords for both object and phenomenon types.
The number of data instances in each dataset is 300, and the average input lengths of failure detail, cause, and measurement are 50.05, 129.59, and 87.56, respectively. The average numbers of the keywords in the datasets are 1.00, 1.10, and 1.23. Since the number of data instances is not large, the five-fold cross-validation is used to validate the proposed keyword extractor.

4.2. Implementation Details

The hyper-parameters for RoBERTa follow the settings of klue/bert-base (https://huggingface.co/klue/bert-base, accessed on 18 October 2023). The embedding size of hidden state is 768, the number of attention heads and the number of layers are twelve, and the size of vocabulary is 32,000. The cross-entropy is adopted for loss function to train the proposed keyword extractor, and AdamW is adopted as an optimizer with a learning rate of 2 × 10 8   β 1 , β 2 , and ϵ for AdamW are set as 0.9, 0.999, and 1 × 10 8 , respectively.
In order to implement the proposed keyword extractor, Python 3.9.17, PyTorch 2.0, and Transformers 4.33.1 are used. The environment for training the keyword extractor is a PC with an RTX 3090-Ti GPU, an Intel i9-10980XE CPU, and 256GB RAM. The maximum sequence length of an input is 256, and the size of mini-batch is 16.

4.3. Evaluation Metrics

The performance of the proposed keyword extractor is measured with well-known automatic metrics: F1-Score, Micro (micro-accuracy), and Macro (macro-accuracy). F1-Score calculates the harmonic mean of the precision and the recall, while Micro calculates the token-level accuracy and Macro calculates the instance-level accuracy.

4.4. Baselines

The proposed keyword extractor is compared with the following baselines:
  • VR: a vanilla RoBERTa model with a basic sequence labeling.
  • BIO: a sequence labeling RoBERTa model with the BIO-tags.
  • SQUAD: a RoBERTa model with SQuAD tags used in SQuAD dataset [41].
  • TNT-KID: a modified transformer-based keyword extractor that additionally incorporates part-of-speech information [10].
  • BART: a transformer-based generative language model.
VR, BIO, and SQUAD are all RoBERTa models, while BART is a standard BART [26]. However, VR, BIO, and SQUAD are differently trained because the instances are labeled in different schemes. In VR, all token labels are assumed to be independent from one another. Thus, no word-level information is used by VR. On the other hand, the keywords are labeled with the BIO-tags as in NER-tasks in BIO, and they are labeled with a starting index and an ending index as in the SQuAD dataset in SQUAD. Thus, both BIO and SQUAD are trained with some word-level information.

4.5. Experimental Results

Table 2 shows how the post-processing affects the performance of keyword extraction in the object keyword type of the failure detail dataset. In this table, REIN implies the reinforcement post-processing and EXC is the exclusion post-processing. For all metrics, EXC achieves higher performance than REIN consistently. This is because the keyword extraction performance of RoBERTa is sufficiently excellent, and thus predicting a few tokens of a word as keyword can be considered as noise. Therefore, exclusion post-processing is used for all the experiments below.
As explained above, the keywords are tagged for a single data instance from two different points of view. Thus, every instance has two types of keywords. As a result, there can be two ways to determine the keywords of a data instance. One way is to train one model to determine both types, and the other way is to train two distinct models for the two types. For the former, the keywords from both types are all regarded as keywords for a data instance. Table 3 compares these two ways. The failure detail dataset is used to measure the performances in this table. According to this table, ‘Object-Model’ trained only for object type achieves higher performance in object-type keyword extraction than ‘Both-Model’ trained to extract both types of keywords. Similarly, ‘Phen.-Model’ trained only for phenomenon type outperforms ‘Both-Model’ in phenomenon-type keyword extraction. Therefore, it can be inferred that the use of two distinct models is more effective.
Table 4 reports how superior the proposed keyword extractor is to the baselines. The proposed keyword extractor outperforms all baselines in all datasets. The performance differences are all statistically significant with p-value < 0.05 . In particular, TNT-KID and BART show extremely low performances. This is because the number of instances in the datasets is insufficient to newly train the modified transformer or fine-tune BART. That is, BART is ineffective when only small datasets are available. VR, BIO, and SQUAD are differentiated by the label-tagging scheme. SQUAD and BIO have to learn the order of B-tags and I-tags or the keyword positions. Since there are not many data instances to train, learning such information is more difficult than independent keyword prediction. This is why VR is better than both SQUAD and BIO in this table.
The proposed keyword extractor achieves the best performance in all datasets, but Macro-accuracy in failure detail is higher than in cause and measurement. This is because there is only one golden keyword in failure detail, but up to three keywords appear in cause and measurement. Thus, keyword extraction for failure detail is easier than that for other datasets. In addition, SQuAD-tagged classification is intrinsically more difficult than BIO-tagged classification since a classifier has to predict both the starting and ending positions in a SQuAD-tagged dataset, but it determines only whether each word is a keyword or not in a BIO-tagged dataset. Thus, when multiple keywords should be found as in cause and measurement datasets, Macro-accuracy of SQUAD is lower than that of BIO.

4.6. Ablation Study

In Section 3.2, we have discussed the heuristics for two problems of learning with a small number of data instances. Let ‘ONE’ denote the first heuristic that remains at least one keyword. Then, Table 5 shows how effective each heuristic is as an ablation study in failure detail. According to Table 2, ‘EXC’ is superior to ‘REIN’. Thus, only ‘EXC’ is considered for the second heuristic. Whenever each heuristic is removed, the performances keep going down. When ‘EXC’ is excluded, Micro- and Macro-Accuracy drop by 1.5 and 6.66, respectively. Moreover, when ‘ONE’ is removed, even F1-Score decreases. In a word, both ‘ONE’ and ‘EXC’ are effective in learning a keyword extractor with a small number of data instances.

5. Conclusions

This paper has proposed a novel keyword extractor trained with an extremely small dataset. Supervised keyword extraction can be divided into classification-based approaches and generation-based approaches. The classification-based approach is formulated as sequence labeling and finds out the keywords of a document by classifying whether each token in the document is a keyword. In contrast, the generation-based approach abstractly generates keywords. Therefore, it is able to generate meaningful keywords that do not appear in the document, but it is based on generative language models, and the generative language models usually require a large amount of data. If a generative language model is trained with a small dataset, it is easy to underfit to the dataset, which leads to poor performance.
The proposed keyword extractor takes the classification-based approach due to the small size of available data. It consists of three components of RoBERTa, a keyword estimator, and a decision rule. The proposed keyword extractor suffers from two problems since it is a token-level classifier based on RoBERTa. The first problem occurs when all tokens in a document are classified as non-keywords, and the second problem is that a single word composed of plural tokens can be determined to be both a keyword and a non-keyword since the token-level classifications are made independently. Therefore, this paper has proposed two kinds of heuristics to solve them. The first heuristic to solve the first problem is to determine the word to which the token with the highest probability belongs as a keyword for the document. The second problem is solved by adopting two post-processing methods: reinforcement and exclusion. The reinforcement post-processing considers a word as a keyword when some tokens of the word are classified as keywords, while the exclusion post-processing determines a word as a keyword only when all tokens of the word are classified as keywords.
The effectiveness of the proposed keyword extractor has been shown through the experiments on Korean Power Plant Outage Reports data. According to the experimental results, the proposed keyword extractor outperforms significantly its baselines in Micro-Accuracy and Macro-Accuracy as well as F1-Score. These results prove that the proposed keyword extractor reduces the problems of a RoBERTa-based keyword extractor trained with a small dataset. In addition, it has also been shown empirically that the exclusion post-processing is superior to the reinforcement post-processing.
One thing to note in the experiments is that BART showed a very low F1-Score. This is because the number of data was so small that BART could not be sufficiently trained. The small data size is also the reason why other RoBERTa-based baselines achieve lower performances than the vanilla RoBERTa. Through an ablation study, it was proven that both the two heuristics are helpful in training the keyword extractor with a small dataset.
Great attention has been paid recently to the generation-based approach in keyword extraction because it enables natural keywords to be generated compared to the classification-based approach [42,43]. As aforementioned above, this approach is difficult to use when the size of available data is small. However, some studies on few-shot learning with a pre-trained language model [44,45] and learning a pre-trained language model for a low-resource language [46,47] are in progress. Thus, it is our future work to combine the proposed keyword extractor with such few-shot learnings.

Author Contributions

Conceptualization, S.-E.K. and J.-B.L.; methodology, S.-E.K.; software, G.-M.P.; validation, J.-B.L. and G.-M.P.; formal analysis, S.-E.K.; resources, S.-B.P.; data curation, S.-M.S.; writing—original draft preparation, S.-E.K.; writing—review and editing, S.-B.P.; visualization, S.-E.K.; supervision, S.-B.P.; project administration, S.-B.P.; funding acquisition, S.-B.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Korea Electric Power Corporation (R22X005-06) and by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2023-RS-2023-00258649) supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation).

Data Availability Statement

Data were obtained from Korea Electric Power Corporation, and the data are not publicly available.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Onan, A. Two-Stage Topic Extraction Model for Bibliometric Data Analysis based on Word Embeddings and Clustering. IEEE Access 2019, 7, 145614–145633. [Google Scholar] [CrossRef]
  2. Bae, Y.S.; Kim, K.H.; Kim, H.K.; Choi, S.W.; Ko, T.; Seo, H.H.; Lee, H.Y.; Jeon, H. Keyword Extraction Algorithm for Classifying Smoking Status from Unstructured Bilingual Electronic Health Records Based on Natural Language Processing. Appl. Sci. 2021, 11, 8812. [Google Scholar] [CrossRef]
  3. Jiang, L.; Meng, Z. Knowledge-Based Visual Question Answering Using Multi-Modal Semantic Graph. Electronics 2023, 12, 1390. [Google Scholar] [CrossRef]
  4. Yuan, Z.; Li, Y.; Li, Y.; Zheng, H.T.; He, Y.; Liu, W.; Huang, D.; Wu, B. A Curriculum Learning Approach for Multi-Domain Text Classification Using Keyword Weight Ranking. Electronics 2023, 12, 3040. [Google Scholar] [CrossRef]
  5. Massaro, A.; Giannone, D.; Birardi, V.; Galiano, A.M. An Innovative Approach for the Evaluation of the Web Page Impact Combining User Experience and Neural Network Score. Future Internet 2021, 13, 145. [Google Scholar] [CrossRef]
  6. Li, M.; Zhang, W.; Hu, B.; Kang, J.; Wang, Y.; Lu, S. Automatic Assessment of Depression and Anxiety through Encoding Pupil-Wave from HCI in VR Scenes. ACM Trans. Multimed. Comput. Commun. Appl. 2023, 20, 42. [Google Scholar] [CrossRef]
  7. Liu, F.; Huang, X.; Huang, W.; Duan, S.X. Performance Evaluation of Keyword Extraction Methods and Visualization for Student Online Comments. Symmetry 2020, 12, 1923. [Google Scholar] [CrossRef]
  8. Campos, R.; Mangaravite, V.; Pasquali, A.; Jorge, A.; Nunes, C.; Jatowt, A. YAKE! Keyword Extraction from Single Documents using Multiple Local Features. Inf. Sci. 2020, 509, 257–289. [Google Scholar] [CrossRef]
  9. da Rocha, N.C.; Barbosa, A.M.P.; Schnr, Y.O.; Machado-Rugolo, J.; de Andrade, L.G.M.; Corrente, J.E.; de Arruda Silveira, L.V. Natural Language Processing to Extract Information from Portuguese-Language Medical Records. Data 2023, 8, 11. [Google Scholar] [CrossRef]
  10. Martinc, M.; Škrlj, B.; Pollak, S. TNT-KID: Transformer-based Neural Tagger for Keyword Identification. Nat. Lang. Eng. 2022, 28, 409–448. [Google Scholar] [CrossRef]
  11. Li, J.; Fan, Q.; Zhang, K. Keyword Extraction based on Tf/Idf for Chinese News Document. Wuhan Univ. J. Nat. Sci. 2007, 12, 917–921. [Google Scholar] [CrossRef]
  12. Wang, J.; Liu, J.; Wang, C. Keyword Extraction based on PageRank. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Nanjing, China, 22–25 May 2007; pp. 857–864. [Google Scholar]
  13. Yao, L.; Pengzhou, Z.; Chi, Z. Research on News Keyword Extraction Technology Based on TF-IDF and TextRank. In Proceedings of the 2019 IEEE/ACIS 18th International Conference on Computer and Information Science, Beijing, China, 17–19 June 2019; pp. 452–455. [Google Scholar]
  14. Kumbhar, A.; Savargaonkar, M.; Nalwaya, A.; Bian, C.; Abouelenien, M. Keyword Extraction Performance Analysis. In Proceedings of the 2019 IEEE Conference on Multimedia Information Processing and Retrieval, San Jose, CA, USA, 28–30 March 2019; pp. 550–553. [Google Scholar]
  15. Li, Y.; Ning, H. Multi-feature Keyword Extraction Method based on TF-IDF and Chinese Grammar Analysis. In Proceedings of the 2021 International Conference on Machine Learning and Intelligent Systems Engineering, Chongqing, China, 9–11 July 2021; pp. 362–365. [Google Scholar]
  16. Yilahun, H.; Hamdulla, A. Entity Extraction based on the Combination of Information Entropy and TF-IDF. Int. J. Reason.-Based Intell. Syst. 2023, 15, 71–78. [Google Scholar] [CrossRef]
  17. Vega-Oliveros, D.; Gomes, P.; Milios, E.; Berton, L. A Multi-centrality Index for Graph-based Keyword Extraction. Inf. Process. Manag. 2019, 56, 102063. [Google Scholar] [CrossRef]
  18. Devika, R.; Subramaniyaswamy, V. A Semantic Graph-based Keyword Extraction Model using Ranking Method on Big Social Data. Wirel. Netw. 2021, 27, 5447–5459. [Google Scholar] [CrossRef]
  19. Kılıç, H.; Çetin, A. A Novel Graph-Based Ensemble Token Classification Model for Keyword Extraction. Arab. J. Sci. Eng. 2023, 48, 10673–10680. [Google Scholar] [CrossRef]
  20. Tang, M.; Gandhi, P.; Kabir, A.; Zou, C.; Blakey, J.; Luo, X. Progress Notes Classification and Keyword Extraction Using Attention-Based Deep Learning Models with BERT. arXiv 2019, arXiv:1910.05786. [Google Scholar]
  21. Grootendorst, M. KeyBERT: Minimal Keyword Extraction with BERT. Available online: https://zenodo.org/records/4461265 (accessed on 18 October 2023).
  22. Lim, Y.; Seo, D.; Jung, Y. Fine-tuning BERT Models for Keyphrase Extraction in Scientific Articles. J. Adv. Inf. Technol. Converg. 2020, 10, 45–56. [Google Scholar] [CrossRef]
  23. Qian, Y.; Jia, C.; Liu, Y. Bert-Based Text Keyword Extraction. J. Phys. Conf. Ser. 2021, 1992, 042077. [Google Scholar] [CrossRef]
  24. Kulkarni, M.; Mahata, D.; Arora, R.; Bhowmik, R. Learning Rich Representation of Keyphrases from Text. arXiv 2021, arXiv:2112.08547. [Google Scholar]
  25. Glazkova, A.; Morozov, D. Multi-Task Fine-tuning for Generating Keyphrases in a Scientific Domain. In Proceedings of the 2023 IX International Conference on Information Technology and Nanotechnology, Samara, Russia, 17–21 April 2023; pp. 1–5. [Google Scholar]
  26. Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7871–7880. [Google Scholar]
  27. Chen, H.; Wang, T.; Chen, T.; Deng, W. Hyperspectral Image Classification Based on Fusing S3-PCA, 2D-SSA and Random Patch Network. Remote Sens. 2023, 15, 3402. [Google Scholar] [CrossRef]
  28. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A Robustly Optimized Bert Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  29. Sparck, K. A Statistical Interpretation of Term Specificity and Its Application in Retrieval. In Document Retrieval Systems; Emerald Group Publishing: Bingley, UK, 1988; pp. 132–142. [Google Scholar]
  30. El-Beltagy, S.; Rafea, A. KP-Miner: A Keyphrase Extraction System for English and Arabic Documents. Inf. Syst. 2009, 34, 132–144. [Google Scholar] [CrossRef]
  31. Rose, S.; Engel, D.; Cramer, N.; Cowley, W. Automatic Keyword Extraction from Individual Documents; Wiley & Sons: Hoboken, NJ, USA, 2010. [Google Scholar]
  32. Mihalcea, R.; Tarau, P. TextRank: Bringing Order into Text. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; pp. 404–411. [Google Scholar]
  33. Danesh, S.; Sumner, T.; Martin, J. SGRank: Combining Statistical and Graphical Methods to Improve the State of the Art in Unsupervised Keyphrase Extraction. In Proceedings of the Fourth Joint Conference on Lexical and Computational Semantics, Denver, CO, USA, 4–5 June 2015; pp. 117–126. [Google Scholar]
  34. Yang, S.; Lu, W.; Yang, D.; Li, X.; Wu, C.; Wei, B. KeyphraseDS: Automatic Generation of Survey by Exploiting Keyphrase Information. Neurocomputing 2017, 224, 58–70. [Google Scholar] [CrossRef]
  35. Luan, Y.; Ostendorf, M.; Hajishirzi, H. Scientific Information Extraction with Semi-supervised Neural Tagging. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 2641–2651. [Google Scholar]
  36. Chun-Wei, J.; Shao, Y.; Djenouri, Y.; Yun, U. ASRNN: A Recurrent Neural Network with an Attention Model for Sequence Labeling. Knowl.-Based Syst. 2021, 212, 106548. [Google Scholar] [CrossRef]
  37. Meng, R.; Yuan, X.; Wang, T.; Brusilovsky, P.; Trischler, A.; He, D. Does Order Matter? An Empirical Study on Generating Multiple Keyphrases as a Sequence. arXiv 2019, arXiv:1909.03590. [Google Scholar]
  38. Wu, H.; Ma, B.; Liu, W.; Chen, T.; Nie, D. Fast and Constrained Absent Keyphrase Generation by Prompt-Based Learning. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; pp. 11495–11503. [Google Scholar]
  39. Bujel, K.; Yannakoudakis, H.; Rei, M. Zero-Shot Sequence Labeling for Transformer-Based Sentence Classifiers. In Proceedings of the 6th Workshop on Representation Learning for NLP, Online, 6 August 2021; pp. 195–205. [Google Scholar]
  40. Yasuoka, K. Sequence-Labeling RoBERTa Model for Dependency-Parsing in Classical Chinese and Its Application to Vietnamese and Thai. In Proceedings of the 2023 8th International Conference on Business and Industrial Research, Bangkok, Thailand, 18–19 May 2023; pp. 169–173. [Google Scholar]
  41. Rajpurkar, P.; Zhang, J.; Lopyrev, K.; Liang, P. SQuAD: 100,000+ Questions for Machine Comprehension of Text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 2383–2392. [Google Scholar]
  42. Piedboeuf, F.; Langlais, P. A New Dataset for Multilingual Keyphrase Generation. In Proceedings of the Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; pp. 38046–38059. [Google Scholar]
  43. Diya, A.; Mizuho, I. Keyphrase Generation by Utilizing BART Finetuning and BERT-Based Ranking. In Proceedings of the Forum on Data Engineering and Information Management, Online, 27 February–2 March 2022. [Google Scholar]
  44. Beltagy, I.; Cohan, A.; Logan, R., IV; Min, S.; Singh, S. Zero-and Few-shot NLP with Pretrained Language Models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts, Dublin, Ireland, 22–27 May 2022; pp. 32–37. [Google Scholar]
  45. Alayrac, J.B.; Donahue, J.; Luc, P.; Miech, A.; Barr, I.; Hasson, Y.; Lenc, K.; Mensch, A.; Millican, K.; Reynolds, M.; et al. Flamingo: A Visual Language Model for Few-Shot Learning. In Proceedings of the Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; pp. 23716–23736. [Google Scholar]
  46. Ogundepo, O.; Oladipo, A.; Adeyemi, M.; Ogueji, K.; Lin, J. AfriTeVA: Extending? Small Data? Pretraining Approaches to Sequence-to-sequence Models. In Proceedings of the Third Workshop on Deep Learning for Low-Resource Natural Language Processing, Seattle, WA, USA, 14 July 2022; pp. 126–135. [Google Scholar]
  47. Alabi, J.; Adelani, D.; Mosbach, M.; Klakow, D. Adapting Pre-trained Language Models to African Languages via Multilingual Adaptive Fine-Tuning. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 4336–4349. [Google Scholar]
Figure 1. Overall architecture of sequence-labeling-based keyword extraction using RoBERTa. The example sentence is “GT (Gas Turbine) trip caused by stuck parts…” in Korean, and the “##” in some tokens implies that it is a token that is connected to a previous token without a white space. In this example, ‘GT’ is a keyword because the keyword extraction domain is about finding the objects causing an outage or a failure.
Figure 1. Overall architecture of sequence-labeling-based keyword extraction using RoBERTa. The example sentence is “GT (Gas Turbine) trip caused by stuck parts…” in Korean, and the “##” in some tokens implies that it is a token that is connected to a previous token without a white space. In this example, ‘GT’ is a keyword because the keyword extraction domain is about finding the objects causing an outage or a failure.
Electronics 12 04560 g001
Figure 2. The examples of two types of post-processing. The example in (a) is “Trip estimation due to vibration sensor malfunction”, and the example in (b) is “Damage of air supply pipe for blow-off V/V operation”. If the label is 1, it is classified as a keyword, and, if the label is 0, it is classified as a non-keyword. Words are separated through wide spacing.
Figure 2. The examples of two types of post-processing. The example in (a) is “Trip estimation due to vibration sensor malfunction”, and the example in (b) is “Damage of air supply pipe for blow-off V/V operation”. If the label is 1, it is classified as a keyword, and, if the label is 0, it is classified as a non-keyword. Words are separated through wide spacing.
Electronics 12 04560 g002
Table 1. An example of Korean Power Plant Outage Reports dataset.
Table 1. An example of Korean Power Plant Outage Reports dataset.
DatasetDocumentObjectPhenomenon
failure detailST 복수기 냉각수(해수) 펌프 고장으로 인한 ST TripSTTrip
ST Trip due to ST condenser coolant (sea water) pump failureSTTrip
causem ST 복수기 냉각수(해수)펌프 고장에 의한 냉각수 공급중단으로 복수기 진공도 하락 후 ST 진동 상승하여 TripST 복수기 냉각수(해수)펌프고장
m ST condenser vacuum level drops due to cooling water supply interruption due to failure of the ST condenser coolant (seawater) pump, then ST vibration increases and tripsST condenser coolant (seawater) pumpfailure
measurement냉각수(해수) 공급 펌프 Shaft 등 신품설치 추진냉각수(해수) 공급 펌프 Shaft 등 신품설치 추진신품 설치
Promotion of installing new products such as cooling water (sea water) supply pump shaftcooling water (sea water) supply pumpinstalling new products
Table 2. Comparison of two post-processing methods of reinforcement (REIN) and exclusion (EXC). Scores in bold stand for the leadership among the metrics.
Table 2. Comparison of two post-processing methods of reinforcement (REIN) and exclusion (EXC). Scores in bold stand for the leadership among the metrics.
MetricF1-ScoreMicroMacro
REIN0.6287.2170.00
EXC0.6590.3673.33
Table 3. Comparison of two methods for extracting two types of keywords. ‘Both-Model’ implies a single model trained to extract both types of keywords, while ‘Object-Model’ is a model trained to extract the object-type keywords, and ‘Phen.-Model’ is that to extract the phenomenon-type keywords. Scores in bold stand for the leadership among the keyword types.
Table 3. Comparison of two methods for extracting two types of keywords. ‘Both-Model’ implies a single model trained to extract both types of keywords, while ‘Object-Model’ is a model trained to extract the object-type keywords, and ‘Phen.-Model’ is that to extract the phenomenon-type keywords. Scores in bold stand for the leadership among the keyword types.
Keyword TypeObjectPhenomenon
MetricF1-ScoreMicroMacroF1-ScoreMicroMacro
Both-Model0.5685.2568.330.7587.5971.67
Object-Model0.6590.3673.33---
Phen.-Model---0.8394.5778.33
Table 4. The empirical comparison of the proposed keyword extractor with its baselines. Scores in bold stand for the leadership among the models.
Table 4. The empirical comparison of the proposed keyword extractor with its baselines. Scores in bold stand for the leadership among the models.
Keyword TypeObjectPhenomenon
MetricF1-ScoreMicroMacroF1-ScoreMicroMacro
failure detail
VR0.6485.2361.670.7788.6561.67
BIO0.5380.6745.000.7283.9853.33
SQUAD0.5981.3651.670.7585.2551.67
TNT-KID0.1323.5811.350.1120.0910.13
BART0.09--0.08--
Proposed0.6590.3673.330.8394.5778.33
cause
VR0.5683.1745.000.4787.0841.67
BIO0.5381.1543.330.4185.7736.67
SQUAD0.5080.2541.670.3584.1531.67
TNT-KID0.0818.2509.150.0716.2108.79
BART0.05--0.04--
Proposed0.6293.8453.330.5897.7461.67
measurement
VR0.3683.1533.330.3887.2033.33
BIO0.3482.5133.330.3782.7735.00
SQUAD0.3380.1625.000.3682.1523.33
TNT-KID0.0716.9708.460.0715.2808.63
BART0.03--0.04--
Proposed0.4292.8543.330.4397.0445.00
Table 5. Ablation study on the post-processing methods. Scores in bold stand for the leadership among the metrics.
Table 5. Ablation study on the post-processing methods. Scores in bold stand for the leadership among the metrics.
MetricF1-ScoreMicroMacro
Proposed0.6590.3673.33
– EXC0.6588.8666.67
– ONE0.6485.2361.67
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, S.-E.; Lee, J.-B.; Park, G.-M.; Sohn, S.-M.; Park, S.-B. RoBERTa-Based Keyword Extraction from Small Number of Korean Documents. Electronics 2023, 12, 4560. https://doi.org/10.3390/electronics12224560

AMA Style

Kim S-E, Lee J-B, Park G-M, Sohn S-M, Park S-B. RoBERTa-Based Keyword Extraction from Small Number of Korean Documents. Electronics. 2023; 12(22):4560. https://doi.org/10.3390/electronics12224560

Chicago/Turabian Style

Kim, So-Eon, Jun-Beom Lee, Gyu-Min Park, Seok-Man Sohn, and Seong-Bae Park. 2023. "RoBERTa-Based Keyword Extraction from Small Number of Korean Documents" Electronics 12, no. 22: 4560. https://doi.org/10.3390/electronics12224560

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop