Next Article in Journal
A Multiview Recognition Method of Predefined Objects for Robot Assembly Using Deep Learning and Its Implementation on an FPGA
Next Article in Special Issue
Measuring the Impact of Language Models in Sentiment Analysis for Mexico’s COVID-19 Pandemic
Previous Article in Journal
Implementation of Virtual Training: The Example of a Faculty of Computer Science during COVID-19 for Sustainable Development in Engineering Education
Previous Article in Special Issue
Lexicon-Based vs. Bert-Based Sentiment Analysis: A Comparative Study in Italian
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Emotion Analysis and Dialogue Breakdown Detection in Dialogue of Chat Systems Based on Deep Neural Networks

1
Graduate School of Sciences and Technology for Innovation, Tokushima University, Tokushima 770-8506, Japan
2
Department of Information Engineering, National Institute of Technology, Kagawa College, 551 Kohda, Takuma-cho, Mitoyo-shi 769-1192, Japan
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(5), 695; https://doi.org/10.3390/electronics11050695
Submission received: 29 December 2021 / Revised: 23 February 2022 / Accepted: 23 February 2022 / Published: 24 February 2022
(This article belongs to the Special Issue Deep Learning and Explainability for Sentiment Analysis)

Abstract

:
In dialogues between robots or computers and humans, dialogue breakdown analysis is an important tool for achieving better chat dialogues. Conventional dialogue breakdown detection methods focus on semantic variance. Although these methods can detect dialogue breakdowns based on semantic gaps, they cannot always detect emotional breakdowns in dialogues. In chat dialogue systems, emotions are sometimes included in the utterances of the system when responding to the speaker. In this study, we detect emotions from utterances, analyze emotional changes, and use them as the dialogue breakdown feature. The proposed method estimates emotions by utterance unit and generates features by calculating the similarity of the emotions of the utterance and the emotions that have appeared in prior utterances. We employ deep neural networks using sentence distributed representation vectors as the feature. In an evaluation of experimental results, the proposed method achieved a higher dialogue breakdown detection rate when compared to the method using a sentence distributed representation vectors.

1. Introduction

A number of recently-created chat dialogue systems based on artificial intelligence techniques have the ability to generate flexible response sentences [1,2,3,4]. However, breakdowns often occur during dialogues using these systems. There are, in fact, various types of dialogue breakdowns. Arend et al. [5] investigated breakdowns in human–robot interactions in a case study analyzing dialogue breakdowns in conversations between humans and the robot that was used in the study. Our study targets the detection of dialogue breakdowns by analyzing utterances between dialogue systems such as a chatbot and users of the system.
Many of the methods proposed for dialogue breakdown detection [6,7,8,9,10] are based on word meaning or dialogue acts, changes of topic, etc. Few attempt to use emotional changes or an emotional feature to detect dialogue breakdown, although, in reality, humans are often made to feel uncomfortable, not by a discrepancy in topics, but rather by a lack of consideration of emotional changes or emotional change patterns. The following dialogue sentences can be used to illustrate. ‘S’ indicates an utterance made by a system; ‘U’ indicates an utterance made by a user of the system.
  • S: “Where did you buy those clothes?”
  • U: “I bought them at the ** department store. They are my favorite.”
  • S: “They look cheap.”
The dialogue flow shown here could occur in real life. However, the expression of such a negative opinion of the user’s “favorite clothes” would very likely be taken as an insult and hurt the user’s feelings, causing the dialogue to break down. Although there is no dialogue breakdown in either context or topic here, this sort of dialogue flow should be avoided in order to smoothly move the conversation forward. Moreover, even though the dialogue system in our example may be generally useful, the kind of problem illustrated here may well discourage the user from using the system further in order to avoid being upset by its response.
To conduct a smooth and intimate chat dialogue with a user, it is important for a system to have the ability to agree and sympathize with the user’s utterances according to the user’s emotional changes, in addition to being able to focus on the topics of the dialogue. In this study, we propose a method for detecting dialogue breakdown by comparing emotional tendencies that can be estimated from the utterances of the system and the user. To accomplish our objective, we constructed an emotion level estimator capable of estimating the strength of each emotion based on an utterance distributed representation vector. A dialogue breakdown label is estimated by calculating the similarities among the emotion estimation results obtained by the system from the target utterance, the emotion estimation results from the user’s most recent utterance, and the emotion estimation results by the system from the utterance prior to the user’s most recent utterance; using this emotion estimator, we then use combinations of the similarities as the feature. We also propose another method using feature quantities that are obtained from an emotional level vector based on emotion estimation results from both system and user utterances prior to the target utterances. In an evaluation experiment, we estimate dialogue breakdown labels using the proposed methods and compare the results with those of other methods, including a method based on a similarity vector for a sentence distributed representation and a current state-of-the-art method.
As our experimental dataset, we used the development data of a chat dialogue corpus that was collected in a Project Next NLP dialogue task [11]. This corpus consists of data from recorded dialogues between a chat dialogue system and a user, with annotated dialogue breakdown labels and comments by several annotators [12]. The annotations indicated three types of dialogue breakdowns: “O”, “T”, and “X”; however, detailed categories for each breakdown were not annotated. Higashinaka et al. [13,14] categorized the various kinds of dialogue breakdowns. The meanings of the dialogue breakdown labels are shown below.
  • O: Not a breakdown: it is easy to continue the conversation.
  • T: Possible breakdown: it is difficult to continue the conversation smoothly.
  • X: Breakdown: it is difficult to continue the conversation.
In this study, we focus on breakdowns caused by a lack understanding of emotion, which might correspond to a “lack of sociability” or a “lack of common sense” under the broad breakdown category of “environment” proposed by Higashinaka et al. However, sense-based factors such as emotion are not defined in Higashinaka’s categorizations.
In Section 2, we introduce related research on dialogue breakdown detection, emotion recognition from dialogue text, and utterance intention recognition in dialogue. In Section 3, we describe the proposed emotion estimation method, the extraction of similar pattern features, and an estimation method for dialogue breakdown labels. In Section 4, we present the evaluation experiments, and the results of the proposed emotion estimation method and dialogue breakdown detection method. Section 5 provides a discussion based on the experimental results, whereas Section 6 concludes the paper.

2. Related Works

In this section, we introduce previous research relating to dialogue tearing detection methods, emotion recognition for dialogue texts, the recognition of speech intentions in dialogue, and examples of research related to interaction. Moreover, the relevance and differences between these studies and the present study are discussed.

2.1. Method of Dialogue Breakdown Detection

The baseline system that was distributed in the dialogue breakdown detection challenge [15] uses conditional random fields (CRFs) and considers the labels of the system utterances immediately preceding each system utterance. It takes into account the label of the previous system utterance for each system utterance. The user utterance is assigned the label PREV-O/PREV-T/PREV-X as the dialogue breakdown label of the previous system utterance, which indicates whether the previous system utterance has not broken down, is likely to break down, or is broken down. In the Dialogue Breakdown Detection Challenge, most participating teams used rule-based methods, support vector machines (SVMs), and deep learning methods. In particular, the majority of the participating teams used deep neural network methods.
In the SVM-based method, the word vectors of the system utterance and previous user utterances are used as features. Another method is to learn the word vectors using a recurrent neural network (RNN) or LSTM encoders based on the dialogue behavior of the target sentence and previous utterance as features. Rule-based methods are mainly based on keywords that are extracted by morphological analyzers and exhibit low versatility.
Hori et al. [16] provided an overview of the experimental setup and evaluation results of the 6th Dialogue System Technology Challenge (DSTC6), which aims to develop an end-to-end dialogue system. In it, it is mentioned that Track3’s dialogue breakdown detection technology performed as well as humans in both English and Japanese languages. However, since the dialogue breakdown detection techniques that show the best performance on a particular dataset often end up being specialized for that dataset, there is still room for improvement in terms of generalizability.
Tsunomori et al. [17] selected measures for dialogue breakdown detection in Dialogue Breakdown Detection Challenge 3. In addition to Accuracy, which is the percentage of correct labels, they considered several other measures, including Accuracy (NB, PB + B) and Accuracy (NB + PB, B). As a result of their selection, they concluded that the RSNOD (NB, PB, B) scale is the best, based on the evaluation of the stability of the rankings and the system discrimination. This measure calculates the Root Symmetric Normalized Order-Aware Divergence (RSNOD) for each system utterance from the distribution of the output of the dialogue breakdown detector and the distribution of correct answers.
Using the Nao communication robot, Maitreyee et al. [18] investigated how young and older adults perceive the actions, roles, and goals of dialogue participants in response to dialogue breakdown.
Kontogiorgos et al. [19] analyzed the multimodal behavioral responses of humans to robot dialogue breakdowns in various tasks. In their study, they found that consistent nonverbal behavior is exhibited in response to certain dialogue breakdowns by robots.
Takayama et al. [20] proposed a method for dialogue breakdown detection by training a dialogue breakdown detector based on a set of annotators grouped using clustering and ensembling multiple detectors. Their method performs global and local dialogue breakdown detection using LSTM, and significantly improves the performance of the Dialogue Breakdown Detection Challenge 3 (DBDC3) over the conventional baseline method using conditional probability fields.

2.2. Emotion Recognition from Dialogue Text

Görer [21] proposed an automatic emotion classification method that combines dialogue act modeling and natural language processing approaches, taking into account the temporal flow of the conversation. Our method differs in that we do not model dialogue acts, but use the results of sentence-by-sentence sentiment estimation for dialogue breakdown detection.
Ren et al. [22] proposed a method for correctly recognizing emotions in dialogue by taking into account long-term experience, rich knowledge, and complex patterns of context and emotional states. In their method, they proposed a new concept, the KES model, to enhance semantic information using external knowledge. This model considers both the external knowledge and the semantic role elements of the conversation, and realizes emotion recognition considering both internal and external states.
Chen et al. [23] used a dictionary of emotional expressions for automatic classification on an unlabeled Chinese emotion corpus and used it for supervised learning. They experimented with a text corpus automatically generated using automatic speech recognition (ASR) from speech signals in the Chinese audio–visual database (CHEAVD), and achieved a performance improvement of more than 10% over the baseline. They use oversampling to balance the training of class imbalanced data. In our work, our goal is different in that we aim to improve performance by considering emotion in dialogue breakdown detection rather than the accuracy of emotion estimation results.
Gao et al. [24] constructed a new dataset called Emotional RelAtionship of inTeractiOn (ERATO), which contains multimodal information of video, audio, and text, based on dramas and movies in order to build a model for recognizing emotions from videos. They proposed a model consisting of synchronous modal–temporal attention (SMTA) units for multimodal fusion to perform the task of pairwise emotion relationship recognition (PERR). Since their method proposes a method for recognizing emotions from multimodal information, it is different from our research on textual dialogues.

2.3. Intention or Interaction Recognition in Dialogue

Song et al. [25] proposed a multi-label classification method for recognizing multiple intentions as a method for recognizing speech intentions for concise and non-standardized linguistic expressions. In their study, they reported that BERT features achieved the best results in an evaluation experiment using the Chinese Multi-Intentional Dialogue Dataset (CMID-Transportation). If the intentions of the utterances can be correctly estimated, dialogue breakdowns can be detected by detecting utterances that do not match the intentions. However, in the case of Japanese conversation, especially in colloquial speech, the omission of subjects and objects frequently occurs, and polysemous words and expressions with similar meanings, but that need to be used differently depending on the situation, are frequently used. Thus, recognizing multiple speech intentions is a very difficult task. In this study, we aim to detect dialogue breakdown without recognizing speech intentions.
Wei et al. [26] focused on Interaction Style, which classifies the way we interact with others, and proposed a method to classify four types of Interaction Style based on the obtained Interaction Style profile information using Support Vector Machines. Since the method of interaction varies depending on the characteristics of the user and the dialogue system, we believe that building an interaction style recognition model that is specific to each type of interaction would be useful for detecting dialogue breakdowns. However, in our data, the dialogue system is fixed and the user is assumed to be an unspecified person. Therefore, a large amount of dialogue data must be collected and a corpus annotated with interaction styles is required for modeling. In this study, we assume that the discomfort caused by the difference in interaction style causes a change in the user’s emotion, and use the result of emotion estimation as the feature value.
Other methods include Saha et al.’s [27] proposed method for classifying dialogue acts using a multi-task network that takes into account emotions.

3. Proposed Method

The proposed method calculates the similarity between the emotion level vectors estimated for the response sentence of the system and the user utterance, and builds a dialogue breakdown label estimation model using the combinations as similarity pattern features. We consider the situation where the system predicts the dialogue breakdown when it generates the response sentence. Therefore, we use the system’s current (target for breakdown detection) utterance and the system’s and the user’s utterances instead of using the user’s utterance after the system’s current utterance. The flow of the proposed method is illustrated in Figure 1.

3.1. Target Corpus

In this study, we use the dataset that was used in the dialogue breakdown detection challenge, and to which the dialogue breakdown label was assigned, as the target of analysis. This data is a corpus of text recordings of one-on-one dialogues between a human and a dialogue system (chatbot), which were labeled by several third parties as either broken, not broken, or in between. The columns (O,X,T) are the labels of dialogue breakdown. A total of 30 annotators assigned labels for O (not a breakdown), T (possible breakdown), and X (breakdown). For example, if (O,X,T) = (11,9,10), there are 10 annotators who are labeled as broken down. User (U) is a human speaker and has not been given a broken dialogue label, so all values are set to −1. The details of the data are described in detail in the dataset section in Section 4. Table 1 shows the example of the dataset of the dialogue with breakdown labels.

3.2. Emotion Estimation

The pre-processing of emotion estimation requires the construction of a vector generating model that converts sentences, phrases, and words into vectors. To perform this task, we used sentence2vec [28], an implementation of paragraph2vec proposed by Le and Mikolov [29,30]. We adopted the paragraph2vec algorithm, which can express the similarity between paragraphs or sentences by converting texts into a dense vector trained using neural networks. In a study conducted by Le and Mikolov [31], the paragraph2vec algorithm achieved state-of-art results in the sentiment analysis and text classification tasks. Figure 2 shows the framework for training the paragraph vector.

3.2.1. Sentence Embedding

We converted semantically similar/related sentences into similar vectors. The sentence vectors were trained using a vector dimension size of 500 and a context window size of 10. Generally, the larger the vector dimension size, the more detailed the information that can be expressed. Because the context window size represents the number of peripheral context words, the larger the context window size becomes, and the wider the relations that can be considered.
The randomly collected 3,403,658 tweets were tokenized by the Japanese morphological analyzer MeCab [32] and used to create the vector generation model.

3.2.2. Neural Networks

To construct the emotion estimation model, we used an utterance corpus consisting of 42,273 spoken sentences and an emotional expression dictionary containing 19,529 emotional expressions (words or phrases) with annotations indicating emotion type. Based on a systematic chart proposed by Fischer [33], six kinds of emotion labels were annotated to the utterance sentences included in the corpus. Several utterances and emotional expressions were annotated with more than one label. These labels were weighted according to their levels of strength. Table 2 shows the various emotion labels and the number of annotations to our emotional corpus. The example sentences included in the corpus are shown in Table 3. The number of examples, the number of words, and the number of vocabularies are shown in Table 4.
There are many approaches to estimating emotion from sentences using linguistic features [34,35,36,37,38]. These approaches often use machine learning methods such as neural networks for training the emotion estimators [39,40]. Since more than one emotion label can be annotated to the same sentence, our approach also uses neural networks, as they are easy to apply to a multi-label classification task. For training features, we used the sentence distributed representation vector.
In the preprocess for modeling the emotion estimation, the utterance sentences were converted into fixed dimension dense vectors. Before processing, the sentences were morphologically analyzed and split into word units. The vectorization algorithm is based on neural networks. This method is called sentence embedding. We used sentence2vec, which is an implementation of sentence embedding similar to Doc2Vec or Paragraph2Vec. The framework for the training of the paragraph vector is shown in Figure 1.
After vectorizing the data with annotated emotion labels following the method described above, the emotion estimation model was trained with deep neural networks by using the sentence vector data as training data. The structure of the networks is of the fully-connected, four hidden layers, feed-forward type. Parameters such as the unit numbers and the activation function in each layer are shown in Table 5. This emotion estimation model is referred to as “Model-1” in the sections that follow. To train Model-1, we did not use sentences annotated with the “E-1” (Neutral) label. Because a neutral emotion is not distinctive, we believed that it would be difficult to extract significant features. Another type of deep neural network was built, which allowed us to add “Neutral” as emotion category “E-1”. The parameters, unit numbers, and the activation function in each layer are shown in Table 6. Hereinafter, this second emotion estimation model is referred to as “Model-2”. Each layer is defined as a fully-connected layer. According to the results from several tests, the two models with different architectures achieved the best scores. Therefore, we decided to use these architectures to estimate emotion from the sentences. The parameter settings for these models, such as the number of units, activation function, and dropout rate, were determined by selecting the best-performing model through training and testing on a randomly selected supervised corpus. Categorical cross-entropy was used as the loss function and Adam was used as the optimization algorithm. For the network configuration of the proposed model and the combinations of parameters, we attempted several patterns and selected the one that resulted in the best model accuracy.
To avoid the effects of bias in the training data, we applied data augmentation to the sentence distributed representations when we trained Model-2. For each example in the training data, we created 50 augmented data elements by randomly adding noise to the values of the dimensions. We added these data evenly to the training data for each emotion label.

3.3. Extraction of Similarity Pattern Feature

Figure 3 depicts the process of calculating the emotion and sentence similarity vectors from the target utterance and the previous two dialogues. The emotion was estimated based on the sentence vector in the calculation of the emotion similarity vector. The estimation results were output as emotion vectors, and the emotion similarity vector was calculated based on the cosine similarity between each emotion vector. The sentence similarity vectors were calculated directly from the sentence vectors. Breakdowns in the responses of dialogue systems due to a lack of emotional understanding are generally caused by differences between the latest emotion presented in the system’s utterance and the emotion in the user’s latest utterance. Accordingly, the following features were extracted for each system response:
  • Similarity e s i m t 1 , t u , s between the emotion ( E t 1 u ) present in the latest utterance of the user and the emotion ( E t s ) generated in the current system response.
  • Similarity e s i m t 1 , t s , s between the emotion ( E t 1 s ) present in the latest response by the system and the emotion ( E t s ) generated in the current system response.
  • Similarity e s i m t 1 , t 1 u , s between the emotions E t 1 u and E t 1 s .
Emotion ( E t 1 u , E t s , E t 1 s ) indicates the emotion level vector. We hold the values of the top two emotion labels and convert the values of the other emotion labels into zero.
Using a baseline method to compare against the proposed method, the following similarity vector features between the sentence distributed vectors were extracted:
  • Similarity s s i m t 1 , t u , s between the sentence distributed representation ( x t 1 u ) that occurred in the latest user’s utterance and the sentence distributed representation ( x t s ) of the current system response utterance.
  • Similarity s s i m t 1 , t s , s between the sentence distributed representation ( x t 1 s ) that occurred in the previous system response utterance and the sentence distributed representation ( x t s ) of the current system response utterance.
  • Similarity s s i m t 1 , t 1 u , s between the sentence distributed representations S t 1 u and S t 1 s .
Rather than only using the system’s current utterance or the user’s prior utterance, using features calculated from the three utterances indicated in Figure 3 provided the system with more information on the context of the dialogue. The reason that even more utterances were not used is that we believed the system should detect a dialogue breakdown as early in the dialogue as possible and be able to quickly recover.
Table 7 shows an example of dialogue sentences and emotion similarity vectors based on the emotion estimation model using deep neural networks. In the figure, “−1” indicates that there were no annotations, as these lines were the utterance sentences of the user.

3.4. Estimation of Dialogue Breakdown Label

We used canonical correlation analysis, a logistic regression, and recurrent neural networks to train the dialogue breakdown analyzer. Figure 4 shows the creation flow of dialogue breakdown regression model. The following sections describe each method.
Here, we propose a method for dialogue breakdown detection for the current utterance by training a model that simultaneously predicts dialogue breakdown labels for two system utterances using a multi-task learning neural network. It is known that multi-task learning neural networks can construct models efficiently and accurately by constructing networks that can learn multiple prediction targets at the same time and can share features such as relatedness.
In this model, we also propose an improved model in which the vector of emotion estimation results obtained from the emotion estimation model is added to the input.
Figure 5 shows the architecture of the multi-task neural network for dialogue breakdown detection. The input is a vector of distributed representations Xt of two system utterances and one human utterance, and a vector of emotion labels EVt predicted from the system and three human utterances, respectively. The prediction targets are the dialogue breakdown labels Zt−2 and Zt of the two input dialogue system utterances. The emotion vector used as input was that of Model-2.
The parameters of the neural network used in this architecture are shown below.
  • Number of units of each layer
    • Input layers: 500
    • Middle Layers-1: 300, Middle Layers-2: 100
    • Output Layers: 2 × 2 (binary) or 3 × 2 (three-valued)
  • Dropout rate:
    • Middle Layers-1: 0.4, Middle Layers-2: 0.1
  • Activation function: softplus (Middle Layers-1, and 2), softmax (Output layers)
  • Kernel initializer: Glorot’s uniform
  • Batch size: 1024
  • Epochs: 200
Training and testing are performed using 10-fold cross-validation.

3.4.1. Canonical Correlation Analysis

The procedure for estimating the distributions of dialogue breakdown labels, dvi, corresponding to the emotion similarity vectors, evi, can be applied to the problem of estimating other real-valued vectors, dvi. In our study, canonical correlation analysis (CCA) [41,42,43] was used to estimate dialogue breakdowns. Canonical correlation analysis is a method to calculate correlation coefficients for multi-variates. It is also a method to calculate a coordinate system based on the five-fold cross covariance between two datasets.
The method is sometimes used to recommend information by calculating the correlation between different medium types, such as image and text, or music and biological information [44,45]. We used the CCA for estimating dialogue breakdown distribution from the similarity vector based on sentence distributed representation vectors or emotion level vectors.

3.4.2. Logistic Regression

In this study, we created a dialogue breakdown classifier by using logistic regression based on similarity patterns as features. Once the dialogue breakdown dataset was established, we used the dialogue breakdown labels with the maximum frequency as the answer labels, rather than using the distributions of the dialogue breakdown labels. We used logistic regression as a machine learning method, as proposed by Cox [46]. The model is a kind of statistic regression model. It is equivalent to simple perceptron; however, the methods to decide the parameters differ. We used the logistic regression model to train the similarity vectors based on either the sentence distributed representation vectors or the emotion level vectors in order to detect dialogue breakdown.

3.4.3. Recurrent Neural Networks

As our dialogue breakdown detection method using recurrent neural networks, we propose a method using a sequence of sentence vectors or emotion vectors as input features to estimate the dialogue breakdown labels. Three types of recurrent neural networks were used: Recurrent Neural Network (RNN) [47], Long Short-Term Memory (LSTM) [48,49], and Gated Recurrent Unit (GRU) [50]. The architectures of RNN, LSTM, and GRU are shown in Figure 6, Figure 7 and Figure 8. Multi-layer deep neural networks such as RNN have the vanishing gradient problem. LSTM layers have a forget gate. The idea of the forget gate was proposed to solve the vanishing gradient problem. GRU layers have a long short-term memory network without an output gate. Therefore, the learning time is shorter than for the LSTM networks.

4. Experiment and Results

In this section, we describe the experiments that were conducted to evaluate the validity of the proposed method. First, we describe the experimental data, experimental conditions, and evaluation method. Thereafter, we present the results of the evaluation experiments of the emotion estimation method and dialogue breakdown detection.

4.1. Experimental Data

We used the development (dev) dataset of Dialogue Breakdown Detection Challenge 2 as the experimental data. The dataset includes the system–user dialogue by following three types of dialogue system.
  • DCM: chat dialogue system using NTT Docomo Chat Dialogue API [51];
  • DIT: chat dialogue system by DENSO IT Laboratory inc. [52];
  • IRS: example-based chat dialogue system based on IR-STATUS [53].
The numbers of dialogues and utterances for each system are shown in Table 8. We used the baseline method using sentence distributed vector as feature as a comparison target.

4.2. Evaluation of Emotion Estimation

To evaluate the performance of the two emotion estimation models, we compared the emotion estimation results from the two models with human judgment. Table 9 shows the number of participants in our test, along with their ages. Using more than 300 utterances (by the systems and the users), we asked each participant to describe the emotion involved by selecting from the list of emotions (joy, anger, sorrow, etc.) shown in Table 1. The annotation agreement rate was calculated by using the Kappa coefficient value. Based on our calculations, the average Kappa value was 0.282. While this value is not especially high, we judged it not too low on the grounds that it is difficult to match seven kinds of annotation labels.
Normalized label frequency vectors annotated with the participants’ responses were then used as emotion vectors. The precision, recall, and F-value for each label were calculated by comparing the results with the outputs from the two emotion estimation models. The top two emotion labels were then designated as the candidates for the output and the answer. Precision, recall, and F-value were used as the criteria in our evaluation. Table 10 presents the distribution of the emotion tags in the dialogue corpus used in the evaluation.
The experimental results are presented in Table 11 and Table 12. As shown, the Model-2 results include the added E-1 (“neutral”) emotion category; however, because Model-1 was trained without using the E-1 label, this category does not appear in the Model-1 results.
As can be seen, the average F-score for Model-1 was higher than that of Model-2; however, the B-1 values in Model-1 could not be estimated, since the results from the participant questionnaires included a large number of E-1 (“neutral”) labels. Model-2 included “calm” as an output label type and could distinguish the types of emotions in more detail than Model-1. The fact that the dropout rate between the hidden layers was lower, and that the activation function was not used in hidden layers 3 and 4, are considered to be the reasons that Model-2 could produce relatively complete emotion estimation results. However, because the F-values were generally low and biased, it is difficult to determine which performance is better.
Based on these results, Model-2 could estimate the E-1 (“neutral”) labels with 80.2% precision. In fact, because many utterances did not express any particular emotion, Model-2, which could estimate “neutral”, would appear to be generally more effective than Model-1 in estimating the emotion in utterances for the dialogue system.

4.3. Evaluation of Dialogue Breakdown Detection

To evaluate the effectiveness of the proposed method, we conducted an evaluation experiment. Dialogue data (50 pairs of dialogue data for development) were obtained from the three types of chat dialogue systems (DCM, DIT, IRS) and used as the dataset for Dialogue Breakdown Detection Challenge 2. We sorted the data according to each system and used the different datasets for the experiment.
The development data were only used to evaluate the data in a 50-fold cross-validation test by each dialogue unit. We compared the proposed method and the baseline method (s2v: using the similarity vector of the sentence distributed representation vector as the feature) using the following evaluation procedure:
  • For the distribution of correct labels and the distribution of the labels that were outputted by the dialogue breakdown, compare
    Cosine Similarity,
    Jensen-Shannon Divergence (JSD) [54],
    Mean Squared Error (MSE).
  • Apply the detection rate of the dialogue breakdown response sentence (a rate that succeeded in estimating dialogue breakdown sentences correctly) (Accuracy).
Equation (1) shows the cosine similarity between the distribution of P and Q. P is defined as (p1, p2,..., pi,..., pn), and Q is defined as (q1, q2,..., qi,..., qn).
cos ine ( P , Q ) = i = 1 n p i × q i i = 1 n p i 2 i = 1 n q i 2
The Kullback–Leibler divergence [55] between distribution f(x) and g(x) is calculated by Equation (2). The Jensen–Shannon divergence for DJS(P||Q) and DJS(Q||P) is calculated by Equations (3) and (4) using DKL(f(x)||g(x)).
D K L ( f ( x ) | | g ( x ) ) =   f ( x )   log f ( x ) g ( x ) d x
D J S ( P | | Q ) = 1 2 D K L ( P | | 1 2 ( P + Q ) ) + 1 2 D K L ( Q | | 1 2 ( P + Q ) )
D J S ( Q | | P ) = 1 2 D K L ( Q | | 1 2 ( Q + P ) ) + 1 2 D K L ( P | | 1 2 ( Q + P ) )
Mean square error (MSE) is calculated by Equation (5).
MSE = 1 n i = 1 n ( p i q i ) 2
Larger cosine similarity values and higher detection rates, as well as lower JSD and MSE values, indicate better performance. JSD is used to measure the distance between some distributions. The MSE is used to measure the error between the vectors. These measures were also used in the Dialogue Breakdown Detection Challenge as measures of the distributional agreement. The cosine similarity, which is used to calculate the similarity between vectors, was used in this study, as well as the JSD and MSE, to examine the distributional agreement of the output labels.
However, because these indices do not measure the performance of dialogue breakdown detection, it was necessary to evaluate them using values including the accuracy, precision, recall, and F1-score.
The calculation formulas for the accuracy, precision, recall, and F-value (F1-score) are presented in Equations (6)–(9). Cmatch indicates the number of labels output by the classifier that match the correct label. Ctotal indicates the total number of correct labels in the evaluation dataset. The fit rate of label l is denoted as Pl, the recurrence rate is denoted as Rl, and the F-value is denoted as Fl. Ml represents the number of cases in which the correct answer is label l out of the outputs with label l. Ol indicates the number of cases predicted to have the label l. Cl represents the number of cases in which the correct answer label is l.
A c c u r a c y   ( % ) = C m a t c h C t o t a l × 100
P l ( % ) = M l O l × 100
R l ( % ) = M l C l × 100
F l = 2 × P l × R l P l + R l
As described earlier, 50 dialogues obtained from each of three dialogue systems, DCM, DIT, and IRS, were used as the experimental target.
Figure 9 shows the experimental results (Accuracy). In the figure, ‘Model-1′ and ‘Model-2′ indicate the results of the proposed methods, and ‘s2v’ indicates the results of the baseline method. The symbol O means the label “Not a breakdown”, the symbol T means “Possible breakdown”, and the symbol X means “Breakdown”.
For many combinations, better results were obtained by the method using emotion vectors than by the method using sentence vectors (s2v). When we used the similarity-based methods by Logistic Regression or CCA, we could obtain better results than the RNN-based methods (RNN, LSTM, GRU). We then evaluated binary classification using the following two patterns:
  • [OT + X]: O and T are “Not breakdown”, X is “Breakdown”.
  • [O + TX]: O is “Not breakdown”, T and X are “Breakdown.”
These binary classification evaluations are used as the standard evaluation indicators.
Recalls, precision, and F-values were calculated. Figure 10 shows the P–R curve for each result. As can be seen from the graphs, we found that our proposed method using emotion level vectors and neural networks produced better precision for the binary dialogue breakdown label classification task than the baseline method. However, the recalls for our proposed method were lower than those of the baseline method.
The comparisons of breakdown detection accuracy and precision, recalls, and F-values for competition 2 are shown in Table 13, Table 14 and Table 15. In the table, P indicates precision, R indicates recall, and F indicates F-value. In addition, (X) indicates binary classification [OT + X] and (XT) indicates binary classification [O + TX]. The other detection methods for comparison are indicated in Table 16.
As shown, our proposed methods are able to detect [O + TX] labels in the system DCM and DIT with a higher F-value than the other methods. However, the accuracy of our proposed methods was generally lower than the other methods. Consequently, we believe that the emotion level vector is a good feature for detecting dialogue breakdown, even though it is just a simple six- or seven-dimension vector used as a training feature.
Next, we compared the three indices: cosine similarity (cos), Jensen–Shannon divergence (JSD), and mean square error (MSE). Table 17, Table 18 and Table 19 show the comparison of cosine, JSD, and MSE.
Based on these results, we found the three methods (RNN, LSTM, GRU) effective in the distributional matching of the dialogue breakdown labels. When Model-1 was used as the emotion estimator, we obtained better results than for the method using other features, indicating that the emotion vector based on Model-1 is relatively effective in detecting dialogue breakdowns.

5. Discussions

As described in the previous section, our proposed approach using emotion level vectors as a feature produced good results in evaluations based on cosine, JSD, and MSE. However, several methods proposed in the latest dialogue breakdown detection tasks produced better results. Because our proposed method uses emotional information exclusively, it ignores important information such as topics, word similarities, and sentence similarities. It would seem, then, that a method that combines both emotion level vectors and sentence embedding should be developed.
Our proposed method achieved better results than the method using the simple vector feature of sentence embedding. Notably, the precision rate for our proposed method in the binary breakdown label classification task was better than the baseline method. Moreover, based on comparisons with other breakdown detection methods presented in DBDC2, our proposed method worked well in the binary breakdown label detection task [O + TX]. This suggests that the emotional feature is important in the dialogue breakdown detection task. We also believe that our method has the potential to extract important features such as the emotional differences between the system and the user (speaker), or the user’s (speaker’s) negative emotions caused by breakdowns.
The similarity features for the emotion level vectors were not very impressive, largely because the vector dimension was small (only three dimensions), and the similarities were simple values. However, if new features combining sentence similarities and emotion level similarities are applied to the dialogue breakdown detection task, we believe that the approach will produce significant improvements in performance.
Furthermore, the emotion estimation model that was used in this study was not very accurate when annotated limited dialogue data were used as the evaluation set. In the future, it is necessary to improve the accuracy by retraining and transfer learning the emotion estimation model using more accurate vectorization methods such as BERT, or by introducing models such as LSTM and CRF, which can consider the dialogue context and long-term emotions [56,57,58].
The proposed method first uses a model that is trained by a neural network on a supervised corpus with emotion labels of a certain size, and then uses the probability values of the emotion labels that are predicted by the model as emotion vectors. That is, it combines and executes the output results of multiple models in a manner similar to connecting them in a pipeline. Therefore, it is necessary to extract feature vectors from the middle layer to share the parameters of the learned emotion vectors. In the future, we would like to consider sharing parameters by learning multitasking between the estimation of the emotion vectors and the estimation of the dialogue breakdown for efficient performance improvement.

6. Conclusions

In this study, we hypothesized that “lack of emotion recognition”, which is thought to be a factor in dialogue breakdown, would appear in differences of emotion expression tendencies. Therefore, we proposed a method to estimate the distribution of dialogue breakdown labels by extracting similarities between the emotion estimation results of utterances made immediately before the target utterance and using these as a feature.
In the evaluation experiment, we found that the proposed method proved largely superior to the method using the similarity of sentence distributed vectors as the feature. In future work, we intend to investigate whether “lack of emotional understanding” occurs in the dialogue breakdown response sentences. More specifically, we will develop a method to determine how the dialogue system behaves when predicting the other party’s emotions before speaking, and quantitatively evaluate its performance using questionnaires. Furthermore, we would then use this method to analyze the consistency between dialogue breakdown and the presence of emotional understanding.
In analyzing performance, we found that in some instances, having only a few cases had a negative influence on results, as the method was forced to estimate an emotion based on a small number of occurrences of the emotion. In this paper, we used an emotion estimation model that was not specifically designed for dialogues. In the future, we intend to construct a specialized emotion estimation model for a dialogue system where emotion labels are manually annotated. Moreover, we plan to improve detection accuracy by combining our method with the baseline method used in the dialogue breakdown detection challenge or with a topic-similarity-based dialogue. Furthermore, we would construct a dialogue breakdown detector with improved emotional recognition using models that consider context, such as BERT and Transformer, which have been frequently used in recent years.

Author Contributions

Conceptualization, K.M.; data curation, K.M.; funding acquisition, K.M. and M.S.; methodology, K.M.; supervision, K.K. and F.R.; validation, M.S. and M.Y.; visualization, K.M.; writing—original draft, K.M.; writing—review and editing, K.M., M.Y. and F.R. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the 2021 SCAT Research Grant and Grants-in-Aid for Scientific Research KAKENHI JP19K12174 and JP20K12027.

Institutional Review Board Statement

Not applicable.

Acknowledgments

We would like to thank the people who helped us annotate the emotion labels of the dialogues in this study.

Conflicts of Interest

The authors declare no conflict interest.

References

  1. Amazon Alexa. Available online: https://developer.amazon.com/ja/alexa (accessed on 29 December 2021).
  2. Google Assistant. Available online: https://assistant.google.com/intl/ja_jp/ (accessed on 29 December 2021).
  3. Apple Siri. Available online: https://www.apple.com/jp/siri/ (accessed on 29 December 2021).
  4. Microsoft Cortana. Available online: https://www.microsoft.com/en-us/cortana (accessed on 29 December 2021).
  5. Arend, B.; Sunnen, P.; Caire, P. Investigating breakdowns in human robot interaction: A conversation analysis guided single case study of a human-robot communication in a museum environment. Int. J. Mech. Aerosp. Ind. Mechatron. Manuf. Eng. 2017, 11, 5. [Google Scholar]
  6. Higashinaka, R.; Mizukami, M.; Funakoshi, K.; Araki, M.; Tsukahara, H.; Kobayashi, Y. Fatal or not? finding errors that lead to dialogue breakdowns in chat-oriented dialogue systems. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 2243–2248. [Google Scholar]
  7. Takayama, J.; Nomoto, E.; Arase, Y. Dialogue breakdown detection considering annotation biases. In Proceedings of the Dialog System Technology Challenges, DSTC6, Long Beach, CL, USA, 10 December 2017. [Google Scholar]
  8. Sugiyama, H. Dialogue breakdown detection based on estimating appropriateness of topic transition. In Proceedings of the Dialog System Technology Challenges, DSTC6, Long Beach, CL, USA, 10 December 2017. [Google Scholar]
  9. Iki, T.; Saito, A. End-to-end character-level dialogue breakdown detection with external memory models. In Proceedings of the Dialog System Technology Challenges, DSTC6, Long Beach, CL, USA, 10 December 2017. [Google Scholar]
  10. Xie, Z.; Ling, G. Dialogue breakdown detection using hierarchical bi-directional LSTMs. In Proceedings of the Dialog System Technology Challenges, DSTC6, Long Beach, CL, USA, 10 December 2017. [Google Scholar]
  11. Higashinaka, R.; Funakoshi, K.; Kobayashi, Y.; Inaba, M. The dialogue breakdown detection challenge: Task description, datasets, and evaluation metrics. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 2016), Portoroz, Slovenia, 23–28 May 2016; pp. 3146–3150. [Google Scholar]
  12. Dialogue Breakdown Detection Challenge Dataset. Available online: https://sites.google.com/site/dialoguebreakdowndetection2/downloads (accessed on 29 December 2021).
  13. Higashinaka, R.; Funakoshi, K.; Mizukami, M.; Tsukahara, H.; Kobayashi, Y.; Araki, M. Analyzing Dialogue Breakdowns in Chat-oriented Dialogue Systems. In Proceedings of the Interspeech Satelite Workshop, Errors by Humans and Machines in Multimedia, Multimodal and Multilingual Data Processing (ERRARE 2015), Sinaia, Romania, 11–13 September 2015. [Google Scholar]
  14. Higashinaka, R.; Funakoshi, K.; Araki, M.; Tsukahara, H.; Kobayashi, Y.; Mizukami, M. Text Chat Dialogue Corpus Construction and Analysis of Dialogue Breakdown. J. Nat. Lang. Process. 2016, 23, 59–86. (In Japanese) [Google Scholar] [CrossRef] [Green Version]
  15. Higashinaka, R.; Funakoshi, K.; Kobayashi, Y.; Inaba, M. The Dialogue Breakdown Detection Challenge. In Proceedings of the JSAI Technical Report, 75th SIG-SLUD, Tokyo, Japan, 29–30 October 2015. (In Japanese). [Google Scholar] [CrossRef]
  16. Hori, C.; Perez, J.; Higashinaka, R.; Hori, T.; Boureau, Y.-L.; Inaba, M.; Tsunomori, Y.; Takahashi, T.; Yoshino, K.; Kim, S. Overview of the sixth dialog system technology challenge: DSTC6. Comput. Speech Lang. 2019, 55, 1–25. [Google Scholar] [CrossRef]
  17. Tsunomori, Y.; Higashinaka, R.; Takahashi, T.; Inaba, M. Selection of evaluation metrics for dialogue breakdown detection in dialogue breakdown detection challenge 3. Trans. Jpn. Soc. Artif. Intell. 2020, 35, 1–10. (In Japanese) [Google Scholar] [CrossRef]
  18. Maitreyee, M.; Lindgren, H. Younger and older adults perceptions on role, behavior, goal and recovery strategies for managing breakdown situations in human-robot dialogues. In Proceedings of the 9th International User Modeling, Adaptation and Personalization Human-Agent Interaction (HAI 2021), Nagoya, Japan, 9–11 November 2021; pp. 433–437. [Google Scholar] [CrossRef]
  19. Kontogiorgos, D.; Tran, M.; Gustafson, J. A systematic cross-corpus analysis of human reactions to robot conversational failures. In Proceedings of the 23rd ACM International Conference on Multimodal Interaction, ICMI 2021, Montreal, ON, Canada, 18–21 October 2021. [Google Scholar] [CrossRef]
  20. Takayama, J.; Nomoto, E.; Arase, Y. Dialogue breakdown detection robust to variations in annotators and dialogue systems. Comput. Speech Lang. 2019, 54, 31–43. [Google Scholar] [CrossRef]
  21. Görer, B. Emotion classification of text based conversations through dialog act modeling. In Proceedings of the 2014 22nd Signal Processing and Communications Applications Conference (SIU), Trabzon, Turkey, 23–25 April 2014. [Google Scholar] [CrossRef]
  22. Ren, F.; She, T. Utilizing external knowledge to enhance semantics in emotion detection in conversation. IEEE Access 2021, 9, 154947–154956. [Google Scholar] [CrossRef]
  23. Chen, J.; Liu, C.; Li, M. Automatic emotional spoken language text corpus construction from written dialogs in fictions. In Proceedings of the 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017. [Google Scholar] [CrossRef]
  24. Gao, X.; Zhao, Y.; Zhang, J.; Cai, L. Pairwise emotional relationship recognition in drama videos: Dataset and benchmark. In Proceedings of the 29th ACM International Conference on Multimedia (MM 2021), Virtual Online, 20–24 October 2021; pp. 3380–3389. [Google Scholar] [CrossRef]
  25. Song, J.; Luo, Q.; Nie, J. Research and application of multi-round dialogue intent recognition method. In Proceedings of the 2020 16th International Conference on Computational Intelligence and Security (CIS), Guangxi, China, 27–30 November 2020. [Google Scholar] [CrossRef]
  26. Wei, W.-L.; Lin, J.-C.; Wu, C.-H. Interaction style recognition based on multi-layer multi-view profile representation. IEEE Trans. Affect. Comput. 2017, 8, 355–368. [Google Scholar] [CrossRef]
  27. Saha, T.; Gupta, D.; Saha, S.; Bhattacharyya, P. Emotion Aided Dialogue Act Classification for Task-Independent Conversations in a Multi-modal Framework. Cogn. Comput. 2021, 13, 277–289. [Google Scholar] [CrossRef]
  28. Sentence2vec. Available online: https://github.com/klb3713/sentence2vec (accessed on 29 December 2021).
  29. Le, Q.; Mikolov, T. Distributed representations of sentences and documents. In Proceedings of the 31st International Conference on Machine Learning (ICML2014), Beijing, China, 21–26 June 2014; pp. 1188–1196. [Google Scholar]
  30. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. In Proceedings of the CoRR, Athens, Greece, 11–14 September 2013; Available online: http://arxiv.org/abs/1301.3781 (accessed on 29 December 2021).
  31. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed representations of words and phrases and their compositionality. Adv. Neural Inf. Process. Syst. 2013, 26, 3111–3119. [Google Scholar]
  32. MeCab: Yet Another Part-of-Speech and Morphological Analyzer. Available online: http://taku910.github.io/mecab/ (accessed on 29 December 2021).
  33. Fischer, K.W.; Shaver, P.; Carnchan, P. A skill approach to emotional development: From basic- to subordinate-category emotions. In Child Development Today and Tomorrow; Damon, W., Ed.; Wiley: Hoboken, NJ, USA, 1989; pp. 107–136. [Google Scholar]
  34. Poria, S.; Gelbukh, A.; Cambria, E.; Hussain, A.; Huang, G.B. 2014. EmoSenticSpace: A novel framework for affective common-sense reasoning. Knowl. Based Syst. 2014, 69, 108–123. [Google Scholar] [CrossRef] [Green Version]
  35. Ptaszynski, M.; Dybala, P.; Shi, W.; Araki, K. Contextual affect analysis: A system for verification of emotion appropriateness supported with contextual valence shifters. Int. J. Biometrics 2010, 2, 2. [Google Scholar] [CrossRef] [Green Version]
  36. Thelwall, M.; Buckley, K.; Paltoglou, G.; Cai, D. Sentiment strength detection in short informal text. J. Am. Soc. Inf. Sci. Technol. 2010, 61, 2544–2558. [Google Scholar] [CrossRef] [Green Version]
  37. Strapparava, C.; Mihalcea, R. Learning to identify emotions in text. In Proceedings of the 2008 ACM symposium on applied computing, Fortaleza, Brasil, 16–20 March 2008. [Google Scholar]
  38. Liew, S.Y.J.; Turtle, R.H. Exploring fine-grained emotion detection in tweets. In Proceeding of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016. [Google Scholar]
  39. Surace, L.; Patacchiola, M.; Sonmez, B.E.; Spataro, W.; Cangelosi, A. Emotion recognition in the wild using deep neural networks and Bayesian classifiers. In Proceedings of the 19th ACM International Conference on Multimodal Interaction, Glasgow, Scotland, 14–17 November 2017; pp. 593–597. [Google Scholar]
  40. Hassan, A.; Mahmood, A. Deep learning approach for sentiment analysis of short texts. In Proceedings of the 3rd International Conference on Control, Automation and Robotics (ICCAR), Nagoya, Japan, 24–26 April 2017. [Google Scholar]
  41. Fan, X. Canonical correlation analysis as a general analytic model. In Advances in Social Science Methodology; Thompson, B., Ed.; JAI: Greenwich, CT, USA, 1996; Volume 4, pp. 71–94. [Google Scholar]
  42. Fan, X. Canonical correlation analysis and structural equation modeling: What do they have in common? Struct. Equ. Modeling 1997, 4, 65–79. [Google Scholar] [CrossRef]
  43. Thompson, B. Canonical Correlation Analysis: Uses and Interpretation; Sage: Newbury Park, CA, USA, 1984. [Google Scholar]
  44. Akaho, S. A kernel method for canonical correlation analysis. In Proceedings of the International Meeting of Psychometric Society, Osaka, Japan, 15–19 July 2001. [Google Scholar]
  45. Akaho, S. Introduction to canonical correlation analysis: Mutual information extraction from multimodal observations. Brain Neural Netw. 2013, 20, 62–72. (In Japanese) [Google Scholar] [CrossRef] [Green Version]
  46. Cox, D.R. The regression analysis of binary sequences (with discussion). J. Roy. Stat. Soc. B 1958, 20, 215–242. [Google Scholar]
  47. Graves, A.; Liwicki, M.; Fernandez, S.; Bertolami, R.; Bunke, H.; Schmidhuber, J. A novel connectionist system for improved unconstrained handwriting recognition (PDF). IEEE Trans. Pattern Anal. Mach. Intell. 2008, 31, 855–868. [Google Scholar] [CrossRef] [Green Version]
  48. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  49. Gers, A.F.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
  50. Kyunghyun, C.; Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  51. NTT Docomo Chat-Oriented Dialogue API. Available online: https://www.nttdocomo.co.jp/service/developer/smart_phone/analysis/chat/ (accessed on 29 December 2021).
  52. Tsukahara, H.; Uchiumi, K. System utterance generation by label propagation over association graph of words and utterance patterns for open-domain dialogue systems. In Proceedings of the 29th Pacific Asia Conference on Language, Information and Computation (PACLIC 2015), Shanghai, China, 30 October–1 November 2015; pp. 323–331. [Google Scholar]
  53. Ritter, A.; Cherry, C.; Dolan, W.B. Data-driven response generation in social media. In Proceedings of the Empirical Methods in Natural Language Processing (EMNLP 2011), Edinburgh, Scotland, 27–31 July 2011; pp. 583–593. [Google Scholar]
  54. Fuglede, B.; Topsoe, F. Jensen-Shannon Divergence and Hilbert Space Embedding. In Proceedings of the International Symposium on Information Theory, ISIT 2004, Chicago, IL, USA, 27 June–2 July 2004. [Google Scholar]
  55. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  56. Wang, Y.; Zhang, J.; Ma, J.; Wang, S.; Xiao, J. Contextualized Emotion Recognition in Conversation as Sequence Tagging. In Proceedings of the 21th Annual Meeting of the Special Interest Group Discourse and Dialogue, Virtual, 1–3 July 2020; pp. 186–195. [Google Scholar]
  57. Guibon, G.; Labeau, M.; Lefeuvre, L.; Clavel, C. Few-shot emotion recognition in conversation with sequential prototypical networks. Softw. Impacts 2022, 12, 100237. [Google Scholar] [CrossRef]
  58. Poria, S.; Hazarika, D.; Mihalcea, R. DialogueRNN: An Attentive RNN for Emotion Detection in Conversations. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2022; pp. 6818–6825. [Google Scholar]
Figure 1. Flow of proposed method.
Figure 1. Flow of proposed method.
Electronics 11 00695 g001
Figure 2. Framework for training the paragraph vector.
Figure 2. Framework for training the paragraph vector.
Electronics 11 00695 g002
Figure 3. Extraction flow of emotion/sentence similarity vector.
Figure 3. Extraction flow of emotion/sentence similarity vector.
Electronics 11 00695 g003
Figure 4. Creation flow of dialogue breakdown regression model.
Figure 4. Creation flow of dialogue breakdown regression model.
Electronics 11 00695 g004
Figure 5. Multi-task learning of dialogue breakdown label.
Figure 5. Multi-task learning of dialogue breakdown label.
Electronics 11 00695 g005
Figure 6. RNN network using sentence vector or emotion vector as feature.
Figure 6. RNN network using sentence vector or emotion vector as feature.
Electronics 11 00695 g006
Figure 7. LSTM network using sentence vector or emotion vector as feature.
Figure 7. LSTM network using sentence vector or emotion vector as feature.
Electronics 11 00695 g007
Figure 8. GRU network using sentence vector or emotion vector as feature.
Figure 8. GRU network using sentence vector or emotion vector as feature.
Electronics 11 00695 g008
Figure 9. Experimental results (Accuracy).
Figure 9. Experimental results (Accuracy).
Electronics 11 00695 g009
Figure 10. P–R curve of binary classification evaluation results (OT + X) and (O + TX).
Figure 10. P–R curve of binary classification evaluation results (OT + X) and (O + TX).
Electronics 11 00695 g010
Table 1. Example of the dialogue with dialogue breakdown labels.
Table 1. Example of the dialogue with dialogue breakdown labels.
S/UUtterance(O, X, T)
SGood evening. Be careful with heat stroke.(30, 0, 0)
UThank you.(−1, −1, −1)
SOkay.(11, 9, 10)
UHow about that?(−1, −1, −1)
SYou must be careful with heat stroke.(5, 11, 14)
UYes. And I got you a heat stroke at first sight.(−1, −1, −1)
SHeat stroke is okay.(8, 8, 14)
UWhich way are you talking about?(−1, −1, −1)
SPrevention is important.(6, 12, 12)
Table 2. Labels and number of annotations.
Table 2. Labels and number of annotations.
LabelExample of EmotionsNum. of Annotations
A-1Joy, relief, sensation, hope, etc.24,044
A-2Love, respect, like, etc.7124
B-1Surprise, amazement, etc.2350
C-1Anger, hate, spite, etc.13,037
D-1Sorrow, pity, guilt, etc.8201
D-2Anxiety, fear, etc.7816
E-1Neutral8717
Table 3. Example of sentences in the emotion corpus.
Table 3. Example of sentences in the emotion corpus.
TypeExampleEmotion Label
Spoken sentencesYou cheated!C-1
My head may hurt.D-1
Now, I will drink v (≧∇≦) vA-1
hardshipD-1
Emotional words/phrasesbluffC-1
cozyA-1
Table 4. Statistics of the emotion corpus.
Table 4. Statistics of the emotion corpus.
# of Examples61,802
# of Words845,940
# of Vocabularies43,444
Table 5. Network structure and hyper parameters (Model-1).
Table 5. Network structure and hyper parameters (Model-1).
LayerNum. of UnitsActivation FunctionDropout Rate
Input500tanh0.5
Hidden-1128tanh0.5
Hidden-2256tanh0.5
Hidden-3512tanh0.5
Hidden-4256tanh0.5
Output6softmax--
Table 6. Network structure and hyper parameters (Model-2).
Table 6. Network structure and hyper parameters (Model-2).
LayerNum. of UnitsActivation FunctionDropout Rate
Input500tanh0.0
Hidden-1200tanh0.0
Hidden-2500tanh0.1
Hidden-31000--0.1
Hidden-4100--0.1
Output7softmax--
Table 7. Example of dialogue sentences and emotion similarity vectors.
Table 7. Example of dialogue sentences and emotion similarity vectors.
S/UUtteranceEmotion VectorEmotion Similarity (Top 2 Emotions) Vector(O, X, T)
SGood evening. Be careful with heat stroke.A2:0.355 A1:0.330 C1:0.277 E1:0.038-(30, 0, 0)
UThank you.E1:0.965 D1:0.027-(−1, −1, −1)
SOkay.A1:0.482 C1:0.264 A2:0.132 E1:0.1210.00 0.00 0.60(11, 9, 10)
UHow about that?E1:0.980 D1:0.0200.00 0.00 1.00(−1, −1, −1)
SYou must be careful with heat stroke.A1:0.728 A2:0.163 C1:0.088 E1:0.0200.00 0.00 0.86(5, 11, 14)
UYes. And I got you a heat stroke at first sight.E1:0.976 D1:0.0240.00 0.00 1.00(−1, −1, −1)
SHeat stroke is okay.E1:0.900 A1:0.036 C1:0.027 A2:0.026 D1:0.0110.00 1.00 0.04(8, 8, 14)
UWhich way are you talking about?E1:0.951 D1:0.0291.00 1.00 1.00(−1, −1, −1)
SPrevention is important.A1:0.889 A2:0.080 C1:0.0311.00 0.00 0.04(6, 12, 12)
Table 8. Number of dialogues and utterances for each dialogue system.
Table 8. Number of dialogues and utterances for each dialogue system.
System# of
Dialogues
# of Utterances
(System)
# of Utterances
(User)
# of Words
(System)
# of Words
(User)
DCM5055050041974179
DIT50550500150835204
IRS5055050082865341
Table 9. Number of subjects and their attributes.
Table 9. Number of subjects and their attributes.
GenderAgeNumber
Male171
191
209
381
Female202
Table 10. Number of annotated tags for each dialogue system.
Table 10. Number of annotated tags for each dialogue system.
Dialogue SystemSystem/UserEmotion Tag
(Code)
Joy
(A-1)
Love
(A-2)
Surprise
(B-1)
Anger
(C-1)
Sorrow
(D-1)
Anxiety
(D-2)
Neutral
(E-1)
DCMSystem27018877373174320
User212139160447157275
DITSystem216180131145039330
User177149251189140236
IRSSystem214201159239250247
User203172189327825255
Table 11. Experimental results of emotion estimation (Model-1).
Table 11. Experimental results of emotion estimation (Model-1).
EmotionPrecisionRecallF-Value
A-176.581.679.0
A-259.385.169.9
B-10.00.00.0
C-119.160.829.1
D-125.626.225.9
D-217.322.619.6
Average33.046.037.2
Table 12. Experimental results of emotion estimation (Model-2).
Table 12. Experimental results of emotion estimation (Model-2).
EmotionPrecisionRecallF-Value
A-156.060.158.0
A-239.850.344.4
B-125.01.12.1
C-18.663.015.2
D-17.511.59.0
D-26.79.57.8
E-180.240.954.2
Average32.033.827.2
Table 13. Comparison with the other methods presented in DBDC2 (DCM).
Table 13. Comparison with the other methods presented in DBDC2 (DCM).
MethodFeatureAccuracyP(X)R(X)F(X)P(XT)R(XT)F(XT)
RNNs2v37.938.15.69.871.596.982.3
Model-142.948.612.620.072.095.081.9
Model-236.540.04.27.671.396.481.9
LSTMs2v35.347.15.610.071.795.581.9
Model-139.840.96.310.971.497.882.5
Model-238.644.42.85.371.6100.083.5
GRUs2v38.347.67.012.271.296.181.8
Model-142.151.710.517.472.297.883.0
Model-238.450.04.98.971.998.683.2
Multi-task Learnings2v43.036.014.020.064.0100.078.0
s2v_Model-238.024.030.026.064.0100.078.0
HCU_run350.452.029.237.491.039.655.1
smap_run141.543.447.845.573.990.881.5
RSL16BD_run140.50.00.00.0100.00.00.0
NTTCS_run152.747.765.255.184.271.077.0
NTTCS_run256.552.358.455.287.562.472.8
NTTCS_run35247.865.755.383.070.576.2
KIT16_run245.545.357.350.675.382.578.7
OKSAT_run335.538.285.452.873.681.677.4
kanolab_run143.337.292.753.172.790.080.4
Table 14. Comparison with the other methods presented in DBDC2 (DIT).
Table 14. Comparison with the other methods presented in DBDC2 (DIT).
MethodFeatureAccuracyP(X)R(X)F(X)P(XT)R(XT)F(XT)
RNNs2v41.585.646.360.186.699.892.7
Model-129.876.538.050.886.7100.092.9
Model-240.786.647.361.286.7100.092.9
LSTMs2v37.489.746.861.586.6100.092.8
Model-117.773.931.744.486.7100.092.9
Model-233.493.045.461.086.7100.092.9
GRUs2v35.787.246.360.586.6100.092.8
Model-117.777.530.243.586.7100.092.9
Model-232.590.445.960.886.7100.092.9
Multi-task Learnings2v48.011.02.03.077.0100.087.0
s2v_Model-241.023.027.025.077.0100.087.0
HCU_run162.265.281.472.490.174.881.7
HCU_run362.465.581.872.790.484.287.2
smap_run158.459.781.869.084.799.391.4
RSL16BD_run159.154.0100.070.184.3100.091.5
NTTCS_run16461.593.974.489.591.090.3
NTTCS_run265.563.294.375.790.089.189.5
KIT16_run259.159.483.769.586.593.289.7
OKSAT_run158.954.299.270.184.398.890.9
kanolab_run157.153.298.969.182.398.189.5
Table 15. Comparison with the other methods presented in DBDC2 (IRS).
Table 15. Comparison with the other methods presented in DBDC2 (IRS).
MethodFeatureAccuracyP(X)R(X)F(X)P(XT)R(XT)F(XT)
RNNs2v34.154.253.253.773.477.975.6
Model-128.761.528.839.271.6100.083.5
Model-230.750.024.933.271.6100.083.5
LSTMs2v38.553.946.850.172.276.874.4
Model-116.654.324.433.771.6100.083.5
Model-216.651.623.932.771.6100.083.5
GRUs2v34.152.948.850.873.576.575.0
Model-126.156.525.435.071.6100.083.5
Model-216.651.524.933.671.6100.083.5
Multi-task Learnings2v35.017.051.026.067.095.078.0
s2v_Model-236.018.056.027.066.094.077.0
HCU_run153.156.758.957.777.753.863.6
smap_run142.048.354.151.072.598.383.5
RSL16GBD_run255.149.496.165.373.993.082.4
RSL16GBD_run355.349.796.165.574.092.782.3
NTTCS_run258.455.480.165.579.177.378.2
NTTCS_run358.453.984.065.778.982.680.7
KIT16_run249.850.668.858.375.085.780.0
OKSAT_run153.148.788.362.874.086.879.9
OKSAT_run345.150.186.163.475.684.079.6
Table 16. List of the other detection methods for the comparison.
Table 16. List of the other detection methods for the comparison.
NameMethodDifference for Each Run
HCU_run1RNN, Multi-Layer PerceptronMinimize MSE
HCU_run3Minimize MSE and average of four models
smap_run1Neural Conversation Model, SVMUse both outputs of Encoder/Decoder
RSL16BD_run1Word2VecUse breakdown rate of the develop data
RSL16BD_run2Use breakdown rate for each pattern
RSL16BD_run3Combination of run1 and run2
NTTCS_run1Extra Trees RegressionChange features and training data
NTTCS_run2
NTTCS_run3
KIT16_run2Multi-Layer Perceptron, LSTM,
Recurrent Convolutional Neural Networks
Type of dialogue breakdown was not used
OKSAT_run1RuleChange rule
OKSAT_run3
kanolab_run1Word2Vec, RuleLiberalize threshold for judgment
Table 17. Comparison of cosine, JSD, and MSE (DCM).
Table 17. Comparison of cosine, JSD, and MSE (DCM).
MLFeatureCosJSDMSE
CCAs2v0.23061.588113.9775
Model-10.51650.61271.6238
Model-20.53350.64310.5197
LRs2v0.81160.17270.1305
Model-10.81130.17350.1308
Model-20.81010.17350.1310
RNNs2v0.81890.16500.1275
Model-10.82760.15980.1232
Model-20.82210.16300.1261
LSTMs2v0.82290.16280.1260
Model-10.82750.16040.1241
Model-20.82700.16100.1249
GRUs2v0.82400.16230.1256
Model-10.82970.15900.1230
Model-20.82590.16140.1250
Table 18. Comparison of cosine, JSD, and MSE (DIT).
Table 18. Comparison of cosine, JSD, and MSE (DIT).
MLFeatureCosJSDMSE
CCAs2v0.44700.81762.4040
Model-10.57740.46880.8583
Model-20.52420.73800.7088
LRs2v0.85920.13260.1051
Model-10.86030.13260.1048
Model-20.85780.13460.1056
RNNs2v0.87810.12110.0985
Model-10.87500.12470.1014
Model-20.87840.12190.0982
LSTMs2v0.88380.11810.0966
Model-10.87420.12540.1028
Model-20.88040.12120.0983
GRUs2v0.88190.11920.0974
Model-10.87400.12560.1029
Model-20.87980.12120.0982
Table 19. Comparison of cosine, JSD, and MSE (IRS).
Table 19. Comparison of cosine, JSD, and MSE (IRS).
MLFeatureCosJSDMSE
CCAs2v0.26180.96372.7551
Model-10.38120.85842.1685
Model-20.50600.69770.6770
LRs2v0.78280.19540.1375
Model-10.78210.19550.1377
Model-20.78260.19630.1376
RNNs2v0.70280.25930.1773
Model-10.80200.18020.1334
Model-20.79340.18480.1367
LSTMs2v0.72070.24470.1689
Model-10.80770.17750.1317
Model-20.80500.17880.1326
GRUs2v0.70490.25560.1762
Model-10.80690.17770.1318
Model-20.80120.18090.1341
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Matsumoto, K.; Sasayama, M.; Yoshida, M.; Kita, K.; Ren, F. Emotion Analysis and Dialogue Breakdown Detection in Dialogue of Chat Systems Based on Deep Neural Networks. Electronics 2022, 11, 695. https://doi.org/10.3390/electronics11050695

AMA Style

Matsumoto K, Sasayama M, Yoshida M, Kita K, Ren F. Emotion Analysis and Dialogue Breakdown Detection in Dialogue of Chat Systems Based on Deep Neural Networks. Electronics. 2022; 11(5):695. https://doi.org/10.3390/electronics11050695

Chicago/Turabian Style

Matsumoto, Kazuyuki, Manabu Sasayama, Minoru Yoshida, Kenji Kita, and Fuji Ren. 2022. "Emotion Analysis and Dialogue Breakdown Detection in Dialogue of Chat Systems Based on Deep Neural Networks" Electronics 11, no. 5: 695. https://doi.org/10.3390/electronics11050695

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop