Next Article in Journal
Promising Approaches for Heat Utilization in Agricultural Machinery Engines
Previous Article in Journal
Improved Performance of Bifacial Photovoltaic Modules with Low-Temperature Processed Textured Rear Reflector
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Dual-Template Prompted Mutual Learning Generative Model for Implicit Aspect-Based Sentiment Analysis

1
School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
2
Shanghai Key Laboratory of Computer Software Testing and Evaluating, Shanghai 201112, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(19), 8719; https://doi.org/10.3390/app14198719
Submission received: 16 July 2024 / Revised: 30 August 2024 / Accepted: 24 September 2024 / Published: 27 September 2024

Abstract

:
Generative models have shown excellent results in aspect-based sentiment analysis tasks by predicting quadruples by setting specific template formats. The existing research predicts sentiment elements and enhances the dependency between elements using the multi-template prompting method, but it does not realize the information interaction in the generation process, and it ignores the dependency between the prompt template and the aspect terms and opinion terms in the input sequence. In this paper, we propose a Dual-template Prompted Mutual Learning (DPML) generative model to enhance the information interaction between generation modules. Specifically, this paper designs a dual template based on prompt learning and, at the same time, develops a mutual learning information enhancement module to guide each generated training process to interact with iterative information. Secondly, in the decoding stage, a label marking the interactive learning module is added to share the explicit emotional expression in the sequence, which can enhance the ability of the model to capture implicit emotion. On two public datasets, our model achieves an average improvement of 5.3% and 3.4% in F1 score compared with the previous state-of-the-art model. In the implicit sentiment analysis experiment, the F1 score of the proposed model in the data subset containing implicit words is increased by 2.75% and 3.42%, respectively.

1. Introduction

Aspect-Based Sentiment Analysis (ABSA) is an essential field in natural language processing. The ABSA task [1,2] aims to extract Aspect Terms (AT) and Opinion Terms (OT) from a given sentence and analyze the Sentiment Polarity (SP) and Aspect Category (AC) of aspect terms. For example, in the sentence “The computer is beautiful, but the keyboard is uncomfortable.”, “computer” and “keyboard” are aspect terms, while “beautiful” and “uncomfortable” are opinion terms, thus indicating that the sentiment polarities of “computer” and “keyboard” are “positive” and “negative”, respectively. The Aspect–Category–Opinion–Sentiment (ACOS) quadruple extraction task proposed by [3] includes implicit sentiment, allowing for a more precise expression of sentiment. As shown in Figure 1, in restaurant reviews, the sentence “I’ve waited over one hour for food” does not contain aspect terms or opinion terms but expresses negative sentiment towards the restaurant’s service. The quadruple (NULL, service general, NULL, NEG) can be extracted from the sentence.
For the ACOS task, Cai et al. [3] firstly established a series of pipeline baselines to extract quadruples by modeling them step by step through two extraction tasks and two classification tasks. However, this multi-stage propagation approach can lead to severe error propagation. More research has transformed the quadruple prediction task into a text generation task, adopting an end-to-end approach to predict quadruples [4,5,6]. Zhang et al. [2] transformed the quadruple extraction task into a paraphrasing generation problem, where the attributes of elements {(AT, AO, OT, SP)} correspond to the elements { X AT , X AC , X SP , X OT } in a fixed element order filled into templates. This method has become a new paradigm.
Generation-based aspect-level sentiment analysis models inherently impose a sequence on the quadruplet elements despite their inherent unordered nature. Hu et al. [7] experimentally verified that different ordering of elements had different effects on the results. Gou et al. [8] advanced an element order-based prompt learning method considering multiple element orders to build multiple views. However, their methodology employs pretrained models to select an optimal order for input, obtaining a single prediction using different strategies without engaging in interactive template manipulation during the model’s training and inference processes.
Inspired by [7,8], we propose a dual-template prompted mutual learning generative model to solve the ACOS task. Firstly, according to the disorder of quadruple elements, we set two prompt templates, and the elements in each template are marked by specific labels. Then, we construct a T5-based generation framework to generate quadruples. Secondly, to enhance the model’s ability to capture implicit words, we set the enhanced interaction module of marker-Tag, a label classification-enhanced decoding module, in the decoding stage. We add auxiliary suffixes to the input sequence to make the implicit words explicit and introduce the BIO tag, which combines the specific marker features in the template to share the explicit and implicit aspect words and opinion words in the sentence in the decoding stage of the model. In addition, we set up a mutual learning enhancement module to interact with the label information of the two template generation process decoders, make full use of the marker dependence to break the sequential decoding limit of the decoder, and learn the marker information iteratively and interactively.
The main contributions of this paper are as follows:
1.
In the ACOS task, we propose a dual-template prompted quadruple generation model based on the disorder of quadruple elements. Based on marker template features, we develop mutual learning enhancement modules in the decoding process of the dual template generation to interact with the iterative label information.
2.
In order to enhance the ability of the model to capture implicit words, we propose a marker classification-enhanced decoding module, which adds auxiliary fields in the training phase and incorporates label feature information in the decoding phase.
3.
In the ACOS task, the proposed method outperforms the baseline, with F1 scores improved by 5.3% and 3.4%. In the experiments of implicit sentiment analysis, our method improves the F1 score by 2.75% and 3.42%, respectively, in the subset containing implicit terms, demonstrating an excellent ability to capture implicit sentiment.

2. Related Work

2.1. Aspect-Based Sentiment Analysis

Aspect-based sentiment analysis has attracted extensive attention due to its more fine-grained information extraction compared with traditional sentiment analysis. In an ABSA task, the single-element prediction tasks include Aspect Term Extraction (ATE) [9,10], Opinion Term Extraction (OTE) [11,12], Aspect Sentiment Classification (ASC) [13], and Aspect Category Detection (ACD) [14] tasks. In order to perform fine-grained sentiment analysis more comprehensively, researchers have carried out composite tasks to predict multiple elements such as Aspect–Opinion Pair Extraction (AOPE) [15,16], Aspect–Category Sentiment Analysis (ACSA) [17], Aspect–Sentiment Triple Extraction (ASTE) [18,19,20], Aspect–Sentiment Quadruple Extraction (ASQE) [2,21,22,23], etc. Among them, the most challenging task is the ASQE task because it can predict four sentiment elements at once, providing the most complete sentiment analysis results.
For the ASQE task, Zhang et al. [2] transformed the original four-tuple prediction task into a text generation task and obtained the final target sequence by setting the output template of the target. This end-to-end generative modeling forms a new paradigm. Unlike the traditional autoregressive generation mode, LV et al. [6] proposed an efficient hybrid generation framework to generate emotional tuples in parallel and locate the corresponding emotional entities simultaneously. Mao et al. [24] reformulated the ABSA task as a tree sequence-to-path generation problem. Wang et al. [25] addressed the problem of data scarcity in ASQE tasks by realizing data augmentation through the inverse process of text generation quadruples. Ye et al. [21] modeled ASQE as a multi-round machine reading comprehension task and established the association between different subtasks through a multi-turn hierarchical query strategy.
The prompt template of the above generation method is single, and the disorder of elements is not taken into account. Hu et al. [7] believe that the quadruple attribute itself is disordered and found that the template order affects the performance of sequence-to-sequence learning through experiments, and the changes at different positions had different trends. Specifically, they used minimum entropy to select the appropriate template order and used multiple template orders to enhance the data in ASQE tasks to achieve good results. On this basis, Gou et al. [8] combined training and inference with multi-view prompts based on the disorder of quad-element attributes to guide the model to generate multiple sentiment tuples and then vote to select the final result. Their work selects the element order based on the element order using a pre-trained model, and there is no interactive learning for multiple template cue generation processes.
Unlike existing methods, we design a dual-template prompt for the element order of aspect and opinion terms and add a mutual learning module to enhance the interactive training of knowledge information generated through multiple paths.

2.2. Implicit Sentiment Analysis

Most existing tasks focus on learning explicit sentiment elements, and the implicit aspects and opinions in sentences can also fully show the sentiment polarity. Cai et al. [3] constructed two datasets containing implicit sentiment and jointly extracted attribute-opinion in two steps. Then, they predicted category-sentiment, but this pipeline-based method will cause error propagation. Zhou et al. [23] divided the ACOS task into two sub-tasks and adopted a one-step unified solution to solve the aspect category prediction task and the aspect opinion sentiment extraction task at the same time, which could avoid error propagation. Li et al. [26] focused on overlapping quadruples and linearized the quadruples to improve the efficiency of the model. Ju et al. [27] designed input templates to incorporate parts of speech into the model to guide the model to learning implicit emotions. Bao et al. [5] considered the semantic structure of the input sequence, combined syntactic and semantic features to build an opinion tree, and detected all sentiment elements in the opinion tree.
Contrastive learning has been widely adopted in implicit sentiment analysis, demonstrating excellent performance [28,29,30,31]. The contrastive learning approach pulls representations of entities of the same semantic class closer, while representations of entities of different semantic classes are pushed outward. Li et al. [30] utilized supervised contrastive pre-training techniques to experiment on a large-scale sentiment annotated corpus to better capture implicit and explicit sentiment tendencies in reviews. Peper et al. [28] generated negative samples in the decoding stage based on a perturbation mask to make the distance between positive and negative samples longer. Xiong et al. [29] constructed a supervised contrast learning model based on machine reading comprehension. They randomly replaced aspect and opinion terms in the quadruple to obtain negative samples and constructed questions for machine reading comprehension based on positive and negative samples. Xu et al. [31] created informative and adaptive negative samples to enhance data and combined multiple classifiers to extract quadruples.
Inspired by [31], we designed BIO tags, trying to use the decoder to obtain the position information of aspect words and opinion words. At the same time, we added auxiliary suffixes combined with BIO tags to guide the model to learn implicit sentiment.

3. Methodology

In this section, we give the definition of the ACOS task and elaborate on the constituent modules of the DPML model. As shown in Figure 2, it consists of a dual-template prompted generation module, a marker classification-enhanced decoding module, and a marker-based mutual learning-enhanced module.

3.1. Task Definition

Take a sentence W = { w 1 , w 2 , , w n } containing n words. The ACOS task is to predict all sentiment quadruples Q = { ( x AT , x AC , x SP , x OT ) k = 1 } k | Q | containing four elements, the aspect term (AT), sentiment polarity (SP), aspect category (AC), and opinion term (OT), where SP∈{Positive, Neutral, Negative}. They are defined as follows:
  • Aspect terms denote an entity and its aspect, indicating the evaluation target. Terms in a sentence that explicitly refer to an entity are considered explicit aspect terms, such as “computer” and “keyboard” in Figure 1. Implicit aspect terms do not explicitly refer to an entity in the sentence. In the second example in Figure 1, the term is represented as NULL.
  • Sentiment polarity indicates the emotional expression displayed by the sentence, including positive, negative, and neutral.
  • Aspect category represents the unique predefined category for aspects in a specific domain.
  • Opinion terms refer to subjective statements about an aspect, typically subjective words or phrases in the text. Terms in a sentence that explicitly refer to an aspect are considered explicit opinion terms, such as “beautiful” and “uncomfortable” in Figure 1. Implicit opinion terms do not explicitly express an opinion in the sentence, represented as NULL in the second example in Figure 1.

3.2. Dual-Template Prompt Generation Module

To capture the dependencies between emotional elements in a sequence, we introduce a prompt template based on the order of elements. To mark different element types, similar to [7,8], we define the generation prompt template for a sentence as T ACSO : [ AT ] x AT [ AC ] x AC [ SP ] x SP [ OT ] x OT . The markers for x AT , x AC , x SP , x OT are [AT], [AC], [SP], and [OT]. The ACOS task predicts that the resulting quadruples are unordered, but the decoding is sequential.
To enhance the model’s perception of aspect terms and opinion terms, we exchange the positions of x A T and x O T while keeping the positions of other elements unchanged, resulting in another template T OCSA : [ OT ] x OT [ AC ] x AC [ SP ] x SP [ AT ] x AT . To correctly identify the two templates, the prefixes “ACSO:” and “OCSA:” are added to the input. If an input sequence contains multiple quadruples, the special symbol [SSEP] connects them.
We use the pre-trained model T5-base [32] for context encoding based on the transformer architecture [33]. The sentence W is first fed to the T5 encoder to obtain the sentence representation H e . During the decoding stage, the decoding output H d t at the t-th time step is computed based on the previous inputs H d < t and the context input H e , formulated as follows:
H d t = T 5 D e c o d e r ( H d < t , H e )
p θ ( H t + 1 | H e , H d < t ) = s o f t m a x ( W T H d t )
where W is the transition matrix and θ is the initialization hyperparameter.
During the training stage, minimizing the cross-entropy serves as the generation loss:
L G e n = t = 1 N l o g p θ ( H d t | H e , H d < t ) .
where N is the length of the sentence. We construct two prompting templates for two relatively independent generation units, where the prompting templates T A C S O and T O C S A lead to generation losses L G e n A C S O and L G e n O C S A , respectively. The final result is obtained by intersecting the outputs from the dual-template prompting generation model.

3.3. Marker Classification-Enhanced Decoding Module

A sentence contains multiple quadruples with explicit and implicit aspects and opinion terms. Marker-based template methods can recover quadruples using different markers during the inference process. However, the special markers for implicit and explicit sentiments are the same for sequences with implicit sentiments. The generation model fails to distinguish between explicit and implicit sentiment expressions correctly. To address this issue, we design a Marker Classification Enhanced Decoding (MCED) module, as shown in Figure 3. During training, we introduced the auxiliary token “NULL” to the input sequences to help effectively distinguish between implicit aspect terms and opinion words. For sequences containing implicit or explicit aspect terms, we append an explicit suffix “NULL” to obtain the input sequence W = { w 1 , w 2 , , w n 1 , N U L L } , where the sequence length is N. Special markers [AT] and [OT], containing information about aspect terms and opinion terms, are used during the generation process. The designed markers can assist the model in identifying different types of elements and also guide the localization of element features [7,8,34].
We extend the BIO tags by setting nine types of tags, T = { B E A , I E A , B I A , I I A , O , B E O , I E O , B I O , I I O } , which are explicit or implicit aspect words and opinion words in the tag sequence. We utilize the decoding final layer hidden state H d t as the marker representation and compute the label features for aspect terms and opinion terms based on the label position P M .
H M A T = H d t P M AT
H M O T = H d t P M O T
M A T = σ ( W 1 H M A T + b 1 )
M O T = σ ( W 1 H M O T + b 1 )
where W 1 is the transition matrix, b 1 is the bias, and σ denotes the ReLU activation function. The number of quadruples in a sentence is denoted as K, where M A T = { m 1 A T , m 2 A T , , m K A T } and M O T = { m 1 O T , m 2 O T , , m K O T } represent the sets of label features for aspect and opinion terms, respectively.
Aspect terms and opinion terms within multiple quadruples are mutually independent, but aspect terms or opinion terms among themselves can be shared. To predict the BIO tags from the marker features M A T and M O T , using a fully connected network, where m i A T M A T , m i O T M O T and m i A T , m i O T R d × N , we can formulate the process as follows:
p i A T = softmax ( W 2 m i A T + b 2 )
p i O T = softmax ( W 2 m i O T + b 2 )
During the decoding stage, the model can enhance its ability to extract aspect terms and opinion terms by sharing explicit and implicit term features in sequences containing multiple quadruples. This is achieved through the interactive learning between BIO tags and marker features. The training loss function for this process is defined as cross-entropy. Specifically, for the generation processes of prompt templates T A C S O and T O C S A , the corresponding losses are denoted as L M T A C S O and L M T O C S A , respectively, with the following formulas.
L M T A C S O A T = i = 1 K j = 1 N t T ( p i j A T = = t ) · log ( p i j A T | t ) .
L M T A C S O O T = i = 1 K j = 1 N t T ( p i j O T = = t ) · log ( p i j O T | t ) .
L M T A C S O = L M T A C S O A T + L M T A C S O O T

3.4. Marker-Based Mutual Learning Enhancement Module

The elements of the quadruple are inherently unordered. After exchanging the positions of aspect terms and opinion terms, we generated aspect terms and opinion terms with different position information based on dual templates. However, the generation processes of different templates are independent, and the model cannot share the decoded features. We have devised an enhanced training approach based on label template mutual learning to facilitate the information exchange between the dual template generation processes, as shown in Figure 2, where N is the sequence length and H A T , H O T R N × d . The label template features F A T for opinion terms and F O T for aspect terms are obtained through the encoded stages H A T , H O T and the decoding stages HM A T , HM O T , defined as follows:
F A T = σ ( W 3 ( H A T HM A T ) + b 3 )
F O T = σ ( W 3 ( H O T HM O T ) + b 3 )
where W 3 is the weight matrix, b 3 is the bias, and σ is the ReLU activation function. Feeding F A T and F O T into a multi-layer perceptron and considering prompt templates T A C S O and T O C S A , we obtain two decoder’s predictions for aspect terms and opinion terms, respectively, as follows:
y A T A C S O = MLP A C S O ( F A T A C S O )
y O T A C S O = MLP A C S O ( F O T A C S O )
y A T O C S A = MLP O C S A ( F A T O C S A )
y O T O C S A = MLP O C S A ( F O T O C S A )
In this way, the two decoders can interactively learn the dependencies of different template orders. During the training iteration, the two decoders can engage in knowledge learning, mutually reinforcing each other. We define the mutual learning loss for this stage as KL divergence:
L K L A T = KL ( y A T A C S O | | y A T O C S A )
L K L O T = KL ( y O T A C S O | | y O T O C S A )

3.5. Training and Inference

Regarding model training, we employ joint training of dual templates to optimize sequences. The loss function of DPML is as follows:
L = α ( L G e n A C S O + L M T A C S O ) + ( 1 α β ) ( L G e n O C S A + L M T O C S A ) + β ( L K L A T + L K L O T )
In terms of inference, DPML generates two distinct quadruples based on the model trained through prompt learning. To prevent the generated quadruples from deviating from the target template, we employ a Constraint Decoding (CD) strategy to ensure the format’s validity [8]. CD dynamically adjusts the candidate vocabulary during decoding at each time step based on the current token. According to the designed template, the candidate list for “[” should be “AT”, “OT”, “SP”, “AC”. Additionally, to ensure the validity of the content, the emotion element following “[SP]” must be positive, negative, or neutral.
For the two templates T A C S O and T O C S A , we obtain two results Y A C S O and Y O C S A , respectively, from which their linearized versions yield two emotion quadruples, Q A C S O and Q O C S A . The final result Y f i n a l is obtained by taking the intersection of Q A C S O and Q O C S A .

4. Experiment

4.1. Datasets

According to previous research [3,28], we evaluate our method on two benchmark datasets, Restaurant-ACOS and Laptop-ACOS [3], which contain a large number of explicit and implicit aspect terms and opinion terms. The Restaurant-ACOS dataset is constructed based on the SemEval 2016 Restaurant dataset [35] and its extended datasets [12,36]. The Laptop-ACOS dataset was collected from Amazon and includes data on brand-new laptops from the years 2017 and 2018. It covers ten different laptop models from six major brands: MBP, Acer, MSI, Lenovo, ASUS, and Samsung. The data have been annotated based on the explicit and implicit aspects, categories, and sentiment outlined in references [12,36], as well as additional opinion annotations [3]. Multiple PhD students familiar with aspect-based sentiment analysis jointly annotated these four elements [3,37]. Furthermore, following the approach proposed in [2], we include the Rest15 dataset from SemEval [35] to enhance our study on the Aspect Sentiment Quad Prediction (ASQP) task, focusing specifically on aspect-based sentiment quad prediction. This addition enables us to investigate the model’s generalization capabilities in quad-prediction tasks. Consistent with the experimental setup in [3,28], we divide the original dataset into training, validation, and test sets.

4.2. Implementation Details

We adopt the 770M parameter T5-large [32] from the Huggingface Transformers library as our pretrained generative encoder–decoder model. We adopt the AdamW [38] optimizer with a learning rate of 3 × 10 4 to update the model parameters and set the batch size to 16. All experiments were conducted on an NVIDIA RTX 3090 GPU. We use precision (Pre), recall (Rec), and F1 score (F1) as the evaluation metrics.

4.3. Implementation Details for Implicit Sentiment Analysis

To better demonstrate the model’s ability to predict implicit terms, we divide these two datasets into five subsets: EAEO, EAIO, IAEO, IAIO, and Contain-I, following [28,29]. EAEO represents subsets containing both explicit aspect terms and opinion terms, EAIO represents subsets containing explicit aspect terms and implicit opinion terms, IAEO represents subsets containing implicit aspect terms and explicit opinion terms, IAIO represents subsets containing both implicit aspect terms and opinion terms, and Contain-I is the union of EAIO, IAEO, and IAIO, representing subsets containing implicit terms. By partitioning the dataset into five distinct subsets, we can conduct a more granular analysis of the model’s performance across various scenarios. Each subset captures different combinations of explicit and implicit terms, facilitating an evaluation of the model’s ability to handle both types of information. The EAEO subset represents the scenario that is the easiest for the model to handle, where both aspect and opinion terms are explicit. In contrast, the IAIO subset presents the most challenging scenario, where both aspect and opinion terms are implicit. Unlike [28,29], we added a CONTAIN-I subset to investigate how prediction accuracy changes when explicit and implicit information are combined.
Table 1 shows the statistical results of the datasets. In both datasets, the EAEO subset containing explicit aspect terms and opinion terms is the largest. It accounts for 66.40% in the Restaurant-ACOS quadruple type and 56.8% in the Laptop-ACOS, indicating that explicit sentiment expressions are common in both domains. The statistics of the CONTAIN-I dataset show that over 33% of quadruples contain implicit sentiment expressions in Restaurant-ACOS, and over 43% of quadruples do so in Laptop-ACOS. On average, each sentence in the Restaurant-ACOS dataset contains 1.6 quadruples, while each sentence in the Laptop-ACOS dataset contains 1.42 quadruples. Sentences in both datasets often contain multiple quadruples, exhibiting complex sentence structures that increase the complexity of sentiment analysis tasks.

4.4. Compared Methods

To validate the effectiveness of our method, we compare our approach with recent state-of-the-art models.
Extract–classify [3]: This is a two-step execution strategy and the first method to address the ACOS task. It first performs joint extraction of aspect–opinion pairs and then predicts aspect categories and sentiment polarities by establishing a multi-classification model. PARAPHRASE [2]: This transforms the quadruple prediction task into a T5-based semantic generation task according to the specified template. GEN_SCL_NAT [28]: In the generation method based on the T5 model, GEN_SCL_NAT introduces contrastive learning to predict quadruples with implicit terms. It generates negative samples for aspect terms, opinion terms, and sentiment polarity through perturbation to predict implicit emotions. ILO/DLO [25]: This conducts data augmentation based on template order to predict quadruples. ILO/DLO selects the order of elements based on instance-level and dataset-level sorting. MVP [8]: This is a prompt-learning generation model based on the order of elements. It regards one permutation of the quadruple as a view, predicts multiple emotional quadruples, and then aggregates multiple views.

4.5. Effects of the Coefficient

The values of coefficients α and β in Equation (21) are crucial for controlling the relative influence of two different losses. Therefore, we set different values for α and β to explore the impact of these parameters on the ACOS task’s F1 score. First, we fix the value of β at 1, ensuring that the sum of the loss parameters for the two generation templates equals 1. Then, we vary α from 0 to 1 in increments of 0.1 to examine the impact of template switching between aspect and opinion terms in templates T A C S O and T O C S A . As shown in Figure 4, regardless of the dataset (Restaurant-ACOS or Laptop-ACOS), the highest F1 score is achieved when α = 0.5 . This indicates that while the positions of opinion terms and aspect terms differ between the two templates, the semantic information provided by the sequences is equally important.
For the hyperparameter β , we controlled for equal loss parameters of the two templates, such that in Equation (21), α = 1 α β . Then, we varied β from 0 to 0.9 in increments of 0.1. Higher values of β increase the model’s emphasis on interactive learning information, reducing the constraint on the generation process’s loss. As shown in the figure, for both datasets, the highest performance is achieved when α = 0.4 and β = 0.2. This demonstrates that the weighting of the interactive learning templates significantly impacts the experiment. Appropriate interactive learning can aid the information learning process for the two-generation templates. We set α and β to 0.4 and 0.2 based on the above analysis.

4.6. Result Analysis

Table 2 shows the experimental results of our model compared with the baseline models, achieving 65.73% and 47.06%, respectively. According to the observation, the F1 score of our method is better than the baseline model. MVP, ILO, Paraphrase, and other generative models based on t5 model are significantly better than the joint extraction method based on the Bert pre-training model on the ACOS task. MVP, DLO, and ILO all adopt multiple prompt templates. We set the parameters according to the best results in the original paper [8] using five templates for the MVP model and seven templates for DLO and ILO. The more templates are designed, the more paths are generated, and the more time it takes to train. Compared with the generative model with multiple templates, we achieve, on average, 5.3% and 3.4% higher F1 scores on the Restaurant-ACOS and Laptop-ACOS datasets with only two templates. The precision is as high as 71.07% and 52.26%, which is 11.54% and 8.43% higher than other models on average. Compared with F1 and Pre, the Rec value is lower. This is due to the fact that the DMLP model obtains the features of different elements with the help of markers in the template to enhance the interactive iteration of the two template generation processes in the decoding stage.
In addition, we compare the performance of the model on the Rest15 dataset in the ASQP task. The F1 score of DMPL outperforms the existing T5-based model by 4.33% on average, which also shows that DMPL is suitable for all aspect sentiment quadruple prediction tasks.
According to the different types of aspects and opinions, five subsets are constructed by splitting Restaurant-ACOS and Laptop-ACOS. Table 3 presents the F1 scores of our model and the baseline model on the subsets. In this way, the ability of the model to learn implicit words is explored. From Table 3, we can see that the model has the highest quadruple prediction power for the subset of the EAEO type. The EAEO subset only contains explicit aspect terms and opinion terms, while other subsets containing implicit terms pose more challenges to the model. Our method achieves the highest F1 score in the vast majority of subsets. Among them, the datasets containing IA or IO can predict the quadruples more effectively, especially in the Contain-I dataset, which mixes explicit and implicit aspect terms and opinion terms, which shows obvious advantages. The F1 score for the Laptop-ACOS dataset improved by 3.42%, and the F1 score for the Restaurant-ACOS dataset increased by 2.75%. Based on the statistics in Table 1, the IAEO quadruplets in the Laptop-ACOS dataset account for 21.48%, although the EAEO quadruplets, which constitute 56.8% of the total, have a higher F1 score of 58.12%, indicating that the DMPL model excels in predicting implicit opinion terms. In the Restaurant-ACOS dataset, the number of IAEO quadruplets and IAIO quadruplets is approximately 9.5% each, yet the F1 score for the IAIO quadruplets predicted by DMPL is 6.97% higher than that of the IAEO quadruplets. The DMPL model, which incorporates the features of aspect terms and opinion terms during the decoding process, enhances its ability to predict implicit sentiment.

4.7. Ablation Study and Analysis

In order to verify the influence of three key modules in the model, Marker-Based Mutual Learning (MBML)-enhanced training, dual template generation module, and Marker Classification-Enhanced Decoding (MCED) module, we conducted ablation experiments. The results are presented in Table 4.
Firstly, we ablated the MBML module. Compared with the DMPL method, the F1 scores for both datasets decreased by more than 3%. Additionally, in Section 4.4, we explored the hyperparameter β . From Figure 4b, we observe that, compared to the model without the mutual learning enhancement module ( β = 0 ), our model shows performance improvements across all values of β . This indicates that mutual learning enhancement training enables each individual generation module to exchange information, leading to a deeper understanding of the task.
For the MCED module, we first conduct ablation analysis on the Restaurant-ACOS and Laptop-ACOS datasets. The results indicate that the performance degrades when the MCED module is removed, compared to the DPML, which is decreased by 1.52% and 0.47%, respectively. This indicates that the MCED module is able to enhance term feature semantic learning by interacting with aspect terms and opinion terms through BIO tags. Furthermore, to investigate the MCED module’s ability to capturing implicit sentiments, we performed an ablation analysis on subsets focused on implicit sentiment analysis. As shown in Table 5, removing the MCED module resulted in average performance declines of 2.79% and 3.02% on the EAIO and IAIO subsets, respectively. MCED uses “NULL” to express explicit terms, which helps the model to distinguish between the semantics of explicit and implicit terms. Additionally, it uses BIO tags containing location information to interact with feature semantics in the decoding stage, which can significantly enhance the model’s prediction of implicit terms.
For the dual-template prompt generation module, we ablate the T A C S O and T O C S A prompt templates, respectively. Removing either prompt template also eliminates the MBML module. As observed in Table 4, the F1 scores with only the T A C S O or only T O C S A templates are significantly lower than the model without the mutual learning training module. The single-template generation model, such as Paraphrase, makes predictions of the elements of a quadruple based on templates, which to some extent restricts the output positions of the elements, as the decoding process can only proceed linearly. In contrast, the dual-template model is built upon the disorder of the four elements, swapping the positions of aspect terms and opinion terms and setting up two predicted outputs for modeling. The multiple paths of output provide the model with various decoding sequences. This indicates that the dual-template generation mitigates the strong dependency on the quadruple element order during the generation and decoding process. The dual templates also enhance the ability of a single template to capture the unordered characteristics of the quadruples, thereby guiding the model’s generation more effectively.

4.8. Case Study

To verify the effectiveness of DMPL, we compare the prediction results of different methods on a sentence of the test set, as shown in Figure 5. Our model is able to identify all quadruples accurately. In Example 1, extract–classify and MVP have difficulty in the extraction of aspect terms and opinion terms. In Example 2, the extract–classify model has trouble targeting the quadruples (NULL, LAPTOP#QUALITY, negative, terrible). Our model has obvious advantages in sentiment prediction with implicit terms. In addition, the classification of the Laptop-ACOS dataset, such as “SOFTWARE#GENERAL” and “SOFTWARE#GENERAL”, has similar semantics and is diverse, which also brings challenges to the model.

5. Conclusions

This paper presents a dual-template prompt mutual learning enhancement model for the ACOS task. We introduce a dual-template generation method based on prompt learning with element order. During the dual-template generation process, the model incorporates mutual learning to facilitate interactions between the template generation processes for aspect and opinion terms. Additionally, in the decoding phase, the DPML model leverages the MCED module to share multiple quadruple features, thereby enhancing the model’s ability to capture implicit sentiments. Experimental results on benchmark datasets demonstrate that our model outperforms all baseline models, where the F1 score is improved by 5.3% and 3.4% on average.

6. Future Work

Our method focuses on solving the problem of quadruple sentiment prediction. A proposal for future work can be summarized in the following four parts:
1.
The model proposed in this paper can be extended to tuples and triples to achieve a unified framework for aspect-based sentiment analysis [39,40]. Expanding the dataset across different domains and languages with the help of data augmentation techniques will also be considered.
2.
Large Language Models (LLMs) have demonstrated their prowess in sentiment analysis tasks, as evidenced by recent studies. In our work, the T5 model serves as the foundation, with comparisons made against T5-based and BERT-based models. Notably, Ref. [41] reported that zero-shot learning using LLMs was comparable to fine-tuning BERT models. Moving forward, we intend to delve deeper into the application of LLM models to address the challenge of implicit sentiment analysis.
3.
Thirdly, this paper focuses on the same domain as commonly used laptop and restaurant reviews and can further explore cross-domain and cross-language sentiment analysis tasks in the future. Furthermore, the dual templates in this paper are predefined and fixed. To facilitate aspect-based sentiment analysis tasks and to allow for greater flexibility in the influence between different templates, we plan to investigate additional template configurations and dynamic input selection to construct a unified framework for aspect-based sentiment analysis.
4.
This paper also focuses on the ABSA task of text. However, multi-modal aspect-based sentiment analysis fuses information from different modalities, which can better capture emotional elements for sentiment analysis. This will also be the direction of our further exploration in the future.

Author Contributions

Conceptualization, Y.Z.; methodology, Y.Z. and Z.L.; resources, Z.L., Y.Z. and S.C.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z. and Z.L.; visualization, Y.Z.; supervision, S.C. and Z.L.; project administration, Y.Z. and S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the Ministry of Education industry–university cooperative education project No. 231101418285337 and in part by Shanghai University under grant No. 22H00324.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The experimental datasets are publicly available at: https://github.com/NUSTM/ACOS, accessed on 1 August 2021.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, W.; Li, X.; Deng, Y.; Bing, L.; Lam, W. A survey on aspect-based sentiment analysis: Tasks, methods, and challenges. IEEE Trans. Knowl. Data Eng. 2022, 35, 11019–11038. [Google Scholar] [CrossRef]
  2. Zhang, W.; Deng, Y.; Li, X.; Yuan, Y.; Bing, L.; Lam, W. Aspect Sentiment Quad Prediction as Paraphrase Generation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; pp. 9209–9219. [Google Scholar]
  3. Cai, H.; Xia, R.; Yu, J. Aspect-category-opinion-sentiment quadruple extraction with implicit aspects and opinions. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers); Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 340–350. [Google Scholar]
  4. Scaria, K.; Gupta, H.; Goyal, S.; Sawant, S.A.; Mishra, S.; Baral, C. Instructabsa: Instruction learning for aspect based sentiment analysis. arXiv 2023, arXiv:2302.08624. [Google Scholar]
  5. Bao, X.; Wang, Z.; Jiang, X.; Xiao, R.; Li, S. Aspect-based Sentiment Analysis with Opinion Tree Generation. In Proceedings of the IJCAI, Vienna, Austria, 23–29 July 2022; Volume 2022, pp. 4044–4050. [Google Scholar]
  6. Lv, H.; Liu, J.; Wang, H.; Wang, Y.; Luo, J.; Liu, Y. Efficient hybrid generation framework for aspect-based sentiment analysis. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, Dubrovnik, Croatia, 2–6 May 2023; pp. 1007–1018. [Google Scholar]
  7. Hu, M.; Wu, Y.; Gao, H.; Bai, Y.; Zhao, S. Improving Aspect Sentiment Quad Prediction via Template-Order Data Augmentation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 7889–7900. [Google Scholar]
  8. Gou, Z.; Guo, Q.; Yang, Y. MvP: Multi-view Prompting Improves Aspect Sentiment Tuple Prediction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 4380–4397. [Google Scholar]
  9. Xu, H.; Liu, B.; Shu, L.; Philip, S.Y. Double Embeddings and CNN-based Sequence Labeling for Aspect Extraction. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers); Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 592–598. [Google Scholar]
  10. Ma, D.; Li, S.; Wu, F.; Xie, X.; Wang, H. Exploring sequence-to-sequence learning in aspect term extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 3538–3547. [Google Scholar]
  11. Wu, M.; Wang, W.; Pan, S.J. Deep weighted maxsat for aspect-based opinion extraction. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 5618–5628. [Google Scholar]
  12. Fan, Z.; Wu, Z.; Dai, X.; Huang, S.; Chen, J. Target-oriented opinion words extraction with target-fused neural sequence labeling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers); Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 2509–2518. [Google Scholar]
  13. Zhang, X.; Xu, J.; Cai, Y.; Tan, X.; Zhu, C. Detecting Dependency-Related Sentiment Features for Aspect-Level Sentiment Classification. IEEE Trans. Affect. Comput. 2023, 14, 196–210. [Google Scholar] [CrossRef]
  14. Chebolu, S.U.S.; Rosso, P.; Kar, S.; Solorio, T. Survey on aspect category detection. ACM Comput. Surv. 2022, 55, 1–37. [Google Scholar] [CrossRef]
  15. Gao, L.; Wang, Y.; Liu, T.; Wang, J.; Zhang, L.; Liao, J. Question-driven span labeling model for aspect–opinion pair extraction. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 12875–12883. [Google Scholar]
  16. Chen, S.; Liu, J.; Wang, Y.; Zhang, W.; Chi, Z. Synchronous double-channel recurrent network for aspect-opinion pair extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 6515–6524. [Google Scholar]
  17. Bu, J.; Ren, L.; Zheng, S.; Yang, Y.; Wang, J.; Zhang, F.; Wu, W. ASAP: A Chinese Review Dataset Towards Aspect Category Sentiment Analysis and Rating Prediction. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 2069–2079. [Google Scholar]
  18. Liang, S.; Wei, W.; Mao, X.L.; Fu, Y.; Fang, R.; Chen, D. STAGE: Span tagging and greedy inference scheme for aspect sentiment triplet extraction. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 13174–13182. [Google Scholar]
  19. Chen, Y.; Keming, C.; Sun, X.; Zhang, Z. A Span-level Bidirectional Network for Aspect Sentiment Triplet Extraction. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 4300–4309. [Google Scholar]
  20. Chen, S.; Wang, Y.; Liu, J.; Wang, Y. Bidirectional machine reading comprehension for aspect sentiment triplet extraction. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 12666–12674. [Google Scholar]
  21. Ye, S.; Zhai, Z.; Li, R. Enhanced Machine Reading Comprehension Method for Aspect Sentiment Quadruplet Extraction. In ECAI 2023; IOS Press: Amsterdam, The Netherlands, 2023; pp. 2874–2881. [Google Scholar]
  22. Bao, X.; Wang, Z.; Zhou, G. Exploring graph pre-training for aspect-based sentiment analysis. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore, 6–10 December 2023; pp. 3623–3634. [Google Scholar]
  23. Zhou, J.; Yang, H.; He, Y.; Mou, H.; Yang, J. A Unified One-Step Solution for Aspect Sentiment Quad Prediction. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, Singapore, 6–10 December 2023; pp. 12249–12265. [Google Scholar]
  24. Mao, Y.; Shen, Y.; Yang, J.; Zhu, X.; Cai, L. Seq2path: Generating sentiment tuples as paths of a tree. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, 22–27 May 2022; pp. 2215–2225. [Google Scholar]
  25. Wang, A.; Jiang, J.; Ma, Y.; Liu, A.; Okazaki, N. Generative Data Augmentation for Aspect Sentiment Quad Prediction. In Proceedings of the 12th Joint Conference on Lexical and Computational Semantics (* SEM 2023), Toronto, ON, Canada, 13–14 July 2023; pp. 128–140. [Google Scholar]
  26. Li, S.; Zhang, Y.; Lan, Y.; Zhao, H.; Zhao, G. From Implicit to Explicit: A Simple Generative Method for Aspect-Category-Opinion-Sentiment Quadruple Extraction. In Proceedings of the 2023 International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia, 18–23 June 2023; pp. 1–8. [Google Scholar]
  27. Junlang, W.; Xia, L.; Junyi, H.; Yongqiang, Z.; Junteng, M. Enhancing Implicit Sentiment Learning via the Incorporation of Part-of-Speech for Aspect-based Sentiment Analysis. In Proceedings of the 22nd Chinese National Conference on Computational Linguistics, Harbin, China, 3–5 August 2023; pp. 786–800. [Google Scholar]
  28. Peper, J.; Wang, L. Generative Aspect-Based Sentiment Analysis with Contrastive Learning and Expressive Structure. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 6089–6095. [Google Scholar]
  29. Xiong, H.; Yan, Z.; Wu, C.; Lu, G.; Pang, S.; Xue, Y.; Cai, Q. BART-based contrastive and retrospective network for aspect-category-opinion-sentiment quadruple extraction. Int. J. Mach. Learn. Cybern. 2023, 14, 3243–3255. [Google Scholar] [CrossRef]
  30. Li, Z.; Zou, Y.; Zhang, C.; Zhang, Q.; Wei, Z. Learning Implicit Sentiment in Aspect-based Sentiment Analysis with Supervised Contrastive Pre-Training. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; pp. 246–256. [Google Scholar]
  31. Xu, X.; Zhang, J.D.; Xiong, L.; Liu, Z. iACOS: Advancing Implicit Sentiment Extraction with Informative and Adaptive Negative Examples. arXiv 2023, arXiv:2311.03896. [Google Scholar]
  32. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  34. Hu, M.; Bai, Y.; Wu, Y.; Zhang, Z.; Zhang, L.; Gao, H.; Zhao, S.; Huang, M. Uncertainty-Aware Unlikelihood Learning Improves Generative Aspect Sentiment Quad Prediction. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, Singapore, 6–10 December 2023; pp. 13481–13494. [Google Scholar]
  35. Pontiki, M.; Galanis, D.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S.; AL-Smadi, M.; Al-Ayyoub, M.; Zhao, Y.; Qin, B.; De Clercq, O.; et al. Semeval-2016 task 5: Aspect based sentiment analysis. In Proceedings of the ProWorkshop on Semantic Evaluation (SemEval-2016); Association for Computational Linguistics: Stroudsburg, PA, USA, 2016; pp. 19–30. [Google Scholar]
  36. Xu, L.; Li, H.; Lu, W.; Bing, L. Position-Aware Tagging for Aspect Sentiment Triplet Extraction. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 2339–2349. [Google Scholar]
  37. Kim, E.; Klinger, R. Who feels what and why? Annotation of a literature corpus with semantic roles of emotions. In Proceedings of the 27th International Conference on Computational Linguistics, Santa Fe, NM, USA, 20–26 August 2018; pp. 1345–1359. [Google Scholar]
  38. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  39. Gao, T.; Fang, J.; Liu, H.; Liu, Z.; Liu, C.; Liu, P.; Bao, Y.; Yan, W. LEGO-ABSA: A prompt-based task assemblable unified generative framework for multi-task aspect-based sentiment analysis. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 7002–7012. [Google Scholar]
  40. Wang, Z.; Xia, R.; Yu, J. Unifiedabsa: A unified absa framework based on multi-task instruction tuning. arXiv 2022, arXiv:2211.10986. [Google Scholar]
  41. Zhong, Q.; Ding, L.; Liu, J.; Du, B.; Tao, D. Can chatgpt understand too? A comparative study on chatgpt and fine-tuned bert. arXiv 2023, arXiv:2302.10198. [Google Scholar]
Figure 1. Examples ACOS task, where blue denotes aspect terms (AT), orange denotes opinion terms (OT), green denotes aspect categories (AC), and pink denotes sentiment polarity (SP).
Figure 1. Examples ACOS task, where blue denotes aspect terms (AT), orange denotes opinion terms (OT), green denotes aspect categories (AC), and pink denotes sentiment polarity (SP).
Applsci 14 08719 g001
Figure 2. The overall architecture of our dual-template prompted mutual learning (DPML) generation model.
Figure 2. The overall architecture of our dual-template prompted mutual learning (DPML) generation model.
Applsci 14 08719 g002
Figure 3. The process of the marker classification-enhanced decoding module.
Figure 3. The process of the marker classification-enhanced decoding module.
Applsci 14 08719 g003
Figure 4. Effects of α and β on the validation set. Figure (a) shows the changes in F1 scores for the Restaurant-ACOS and Laptop-ACOS datasets as the value of α varies. Panel (b) illustrates the changes in F1 scores for the Restaurant-ACOS and Laptop-ACOS datasets as the value of β varies.
Figure 4. Effects of α and β on the validation set. Figure (a) shows the changes in F1 scores for the Restaurant-ACOS and Laptop-ACOS datasets as the value of α varies. Panel (b) illustrates the changes in F1 scores for the Restaurant-ACOS and Laptop-ACOS datasets as the value of β varies.
Applsci 14 08719 g004
Figure 5. Case study. Example 1 is from the test set of Restaurant-ACOS, and Example 2 is from the test set of Laptop-ACOS.
Figure 5. Case study. Example 1 is from the test set of Restaurant-ACOS, and Example 2 is from the test set of Laptop-ACOS.
Applsci 14 08719 g005
Table 1. Statistical analysis of Restaurant-ACOS and Laptop-ACOS datasets. EA denotes an explicit aspect term, EO denotes an explicit opinion term, IA denotes an implicit aspect term, and IO denotes an implicit opinion term. Over 33% of quadruples in Restaurant-ACOS include implicit sentiment expressions, and over 43% of quadruples in Laptop-ACOS include implicit sentiment expressions.
Table 1. Statistical analysis of Restaurant-ACOS and Laptop-ACOS datasets. EA denotes an explicit aspect term, EO denotes an explicit opinion term, IA denotes an implicit aspect term, and IO denotes an implicit opinion term. Over 33% of quadruples in Restaurant-ACOS include implicit sentiment expressions, and over 43% of quadruples in Laptop-ACOS include implicit sentiment expressions.
Restaurant-ACOSLaptop-ACOS
Categories13121
Sentences22864076
QuadruplesEAEO2429 (66.40%)3269 (56.8%)
EAIO530 (14.49%)910 (15.80%)
IAEO350 (9.57%)1237 (21.48%)
IAIO349 (9.54%)342 (5.94%)
Contain-I1229 (33.6%)2489 (43.22%)
Quadruples/Sentences1.61.42
Table 2. Experimental results on DPML. The best results of previous baselines are bolded.
Table 2. Experimental results on DPML. The best results of previous baselines are bolded.
MethodRestaurant-ACOSLaptop-ACOSRest15
PreRecF1PreRecF1PreRecF1
Extract–classify42.9554.0847.8739.2130.3534.2232.9333.4135.44
Paraphrase59.5159.1159.3143.8742.8143.3349.5348.0648.85
GEN_SCL_NAT58.8859.2759.0742.6542.3942.5247.2247.0747.50
MVP59.5858.7359.1542.8042.6542.7250.6849.0949.87
ILO58.5560.3859.4543.7143.4943.6047.2347.8747.68
DLO60.4760.4060.4344.9344.6244.7748.9248.6948.86
DMPL71.0761.1465.7352.2644.8147.0655.3451.7152.88
Table 3. Implicit sentiment analysis.
Table 3. Implicit sentiment analysis.
MethodRestaurant-ACOSLaptop-ACOS
EAEOEAIOIAEOIAIOContain-IEAEOEAIOIAEOIAIOContain-I
Extract-Classify48.3427.5237.9636.9630.9233.8115.2437.4217.0029.07
Paraphrase64.1840.6151.0252.4347.5545.0935.5152.7435.1542.76
GEN_SCL_NAT63.0638.3745.4855.4547.4443.7431.4452.9433.5440.28
MVP63.8642.5451.0552.7447.6245.4337.0657.0034.1841.95
ILO63.2935.3451.3456.4647.4146.5536.0954.6338.2742.37
DLO63.4838.8950.0056.7145.8745.5837.4455.6437.6342.41
DPML66.4547.8851.7958.7350.3747.5037.4458.1239.4446.18
The best results of previous baselines are bolded.
Table 4. Ablation study.
Table 4. Ablation study.
DatasetMethodPreRecF1
Restaurant-ACOSDLO60.4760.460.43
DPML71.0761.1465.73
w/o MBML72.2558.8464.86
w/o MCED70.3559.0664.21
w/o T O C S A 61.6361.961.76
w/o T A C S O 60.7260.9260.82
Laptop-ACOSDLO44.9344.6244.77
DPML52.2644.8147.06
w/o MBML54.5442.4647.75
w/o MCED52.0941.9546.47
w/o T O C S A 45.0245.1345.08
w/o T A C S O 45.2045.0545.13
Bold numbers represent the best results.
Table 5. Ablation study for implicit sentiment.
Table 5. Ablation study for implicit sentiment.
DatasetMethodEAEOEAIOIAEOIAIOContain-I
Restaurant-ACOSDPML66.4547.8851.7958.7350.37
w/o MCED65.6231.9650.3453.9249.49
Laptop-ACOSDPML47.5037.4458.1239.4446.18
w/o MCED47.235.2556.4436.0042.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lei, Z.; Zhang, Y.; Chen, S. A Dual-Template Prompted Mutual Learning Generative Model for Implicit Aspect-Based Sentiment Analysis. Appl. Sci. 2024, 14, 8719. https://doi.org/10.3390/app14198719

AMA Style

Lei Z, Zhang Y, Chen S. A Dual-Template Prompted Mutual Learning Generative Model for Implicit Aspect-Based Sentiment Analysis. Applied Sciences. 2024; 14(19):8719. https://doi.org/10.3390/app14198719

Chicago/Turabian Style

Lei, Zhou, Yawei Zhang, and Shengbo Chen. 2024. "A Dual-Template Prompted Mutual Learning Generative Model for Implicit Aspect-Based Sentiment Analysis" Applied Sciences 14, no. 19: 8719. https://doi.org/10.3390/app14198719

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop