Next Article in Journal
Trading Community Analysis of Countries’ Roll-On/Roll-Off Shipping Networks Using Fine-Grained Vessel Trajectory Data
Previous Article in Journal
Mangrove Extraction Algorithm Based on Orthogonal Matching Filter-Weighted Least Squares
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Emotion Recognition Method for Humanoid Robot Body Movements Based on a PSO-BP-RMSProp Neural Network

1
Co-Innovation Center of Efficient Processing and Utilization of Forest Resources, Nanjing Forestry University, Nanjing 210037, China
2
College of Furnishings and Industrial Design, Nanjing Forestry University, Nanjing 210037, China
3
The State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(22), 7227; https://doi.org/10.3390/s24227227
Submission received: 6 October 2024 / Revised: 2 November 2024 / Accepted: 8 November 2024 / Published: 12 November 2024
(This article belongs to the Section Sensors and Robotics)

Abstract

:
This paper mainly explores the computational model that connects a robot’s emotional body movements with human emotion to propose an emotion recognition method for humanoid robot body movements. There is sparse research directly carried out from this perspective to recognize robot bodily expression. A robot’s body movements are designed by imitating human emotional body movements. Subjective questionnaires and statistical methods are used to analyze the characteristics of a user’s perceptions and select appropriate designs. An emotional body movement recognition model using a BP neural network (EBMR-BP model) is proposed, in which the selected robot’s body movements and corresponding emotions are used as inputs and outputs. The corresponding topological architecture, encoding rules, and training process are illustrated in detail. Then, the PSO method and the RMSProp algorithm are introduced to optimize the EBMR-BP method, and the PSO-BP-RMSProp model is developed. Through experiments and comparisons for emotion recognition of a robot’s body movements, the feasibility and effectiveness of the EBMR-BP model, with a recognition rate of 66.67%, and the PSO-BP-RMSProp model, with a recognition rate of 88.89%, are verified. This indicates that the proposed method can be used for emotion recognition of a robot’s body movements, and optimization can improve emotion recognition. The contributions are beneficial for emotional interaction design in HRI.

1. Introduction

Emotional interaction in human–robot interaction (HRI) has become an important issue, as humanoid robots are widely applied in different kinds of scenarios like education [1,2], healthcare [3,4,5], shopping [6,7], exhibitions [8], etc. Robots should be capable of communicating emotionally with users. It can make a user’s robot easy to understand and promote interaction in a positive direction. The emotional interaction in HRI can be achieved through facial expressions, voice, body movements, etc. For example, many studies have explored emotional interaction from the perspective of facial expressions, like the impact of affective facial expressions in robots on human behaviors and attitudes [9,10], the estimation and recognition of a user’s emotion [11,12], the generation of robot facial expressions [13,14], etc. There are also many research works referring to affective voices, such as in the evaluation of different robot voices [15,16,17,18], emotional expressions of synthesized speech [19,20,21], etc. Lots of studies focused on combinations of different approaches, such as facial expressions and vocal information, voice and head movements, and so on [22,23,24,25]. However, compared with these mentioned works, there have still been limited studies in the field of robot body movements with emotions up to now, especially for the corresponding emotion recognition.
As body language plays an important role in communicating human emotions in social interactions [25], emotional body movements also have irreplaceable advantages for HRI. They can provide a feasible approach, perceive emotions from a long distance, help identify emotional information when voices are not clear, and also benefit emotion recognition when facial expressions are ambiguous. There have been studies to verify that the recognition accuracy of the affective states expressed by body movements is comparable relative to facial expression [26,27]. Our research focuses on the emotion recognition of a humanoid robot’s body movements during HRI. We hope to explore the relationship between a humanoid robot’s body movements and its emotional state and develop a computational model to recognize the emotions of a humanoid robot’s body movements, which will make the future human-like emotional expression systems of robots in HRI more convenient and natural.
To this end, the design of emotional body movements of robots and the user’s perception of emotions expressed through a robot must be considered. To date, there have been studies working on the design of emotional body movements of robots from different points of view, and the robot platforms Nao, Brian 2.0, Pepper, Teleco, etc., are employed and designed to express different emotions such as joy, anger, shame, sadness, warmth, and so on [28,29,30,31]. However, just like how human body movements are constrained by bones and joints, the implementation of body movements into robots is also constrained by their physical structure. The robot platform we used, a Yanshee robot, has a relatively lower degree of freedom (DoF) compared with the robots mentioned above, which makes it difficult to verify the transferability of body movements designed in other studies. Hence, the design of emotional body movements in this paper is carried out by imitating human emotional body movements, and then they are analyzed to verify the corresponding emotional category.
As mentioned above, to develop the emotion recognition method for a humanoid robot’s body movements during HRI, the work in this paper is described in detail as follows:
  • Construct an effective dataset consisting of the robot’s emotional body movements and corresponding emotional states and explore the trends of user perceptions for different emotions.
  • Propose a computational model to build the mapping between the robot’s emotional body movements and corresponding emotional states and achieve emotion recognition of the robot’s body movements.
  • Verify the effectiveness of the proposed model with experiments on emotion recognition for the emotions of a robot’s body movements and find a feasible approach to optimizing the recognition rate.
The structure of this paper is as follows. Section 2 briefly illustrates the related works, including the design of a robot’s emotional body movements and the corresponding perceptions of users in HRI, as well as the neural network used for emotion recognition in HRI. Section 3 shows the designed robot’s emotional body movements and the implemented procedures and analysis for subjective evaluation. Section 4 proposes an emotion recognition model using a BP neural network for the robot’s body movement. Section 5 illustrates the optimization for the proposed model using PSO and RMSProp. Section 6 illustrates the experiments on emotion recognition through different methods and describes their comparisons. Section 7 shows the conclusions and gives some limitations of this study. The main research procedure of this study is shown in Figure 1. The relationship between Section 3, Section 4, Section 5 and Section 6 is clearly illustrated to improve readability.

2. Related Works

2.1. Design of Robot Emotional Body Movements and Corresponding Perceptions of Users in HRI

Designing a robot’s body movements to accurately express emotions is quite important in the field of robotics. At present, three approaches for this purpose can be summarized. The first is to consider the relationships between the characteristics of body movements and emotions, which can be used to classify different emotional types. For instance, positive and negative emotions can be distinguished by trunk movements, ranging from stretching to bowing [33]; vertical features and features indicating the lateral opening of the body are informative for separating happiness from sadness [34]; and happy and angry expressions usually accompany movements such as bouncing and foot stomping, respectively [35]. Such features have been used in some studies for the design of emotional body movements for robots. Tsiourti et al. employed a Pepper robot and designed body movements reflecting happiness with a straight robot trunk, head bent back, vertical and lateral extension of the arms, and large, fast movements; body movements reflecting surprise with a straight trunk, a backward step, and fast movements; and body movements reflecting sadness with a bowed trunk, downward head, and slow body movements [28]. Fu et al. employed the mobile CommU robot and designed gait-induced vertical oscillations simulating overall body movements to express the emotions of happiness, sadness, and anger [36]. The second approach is to imitate human emotional movements. Robots can express emotions by imitating the motions of users [24,37,38]; for instance, Erden employed a Nao robot and generated the emotional body movements associated with anger, disgust, fear, happiness, sadness, and surprise based on the simplified human body models presented by Coulson [29,39]. The third approach involves designing robot emotional behaviors using motion primitives. Humans can accomplish different tasks by assembling motion primitives (MPs) [40]. Some studies have introduced MPs to express robot motions such as anthropomorphic arms or emotional body movements [40,41,42]. The works of Wei et al. and Li at al. employed Nao robots and provided proof that adopting MPs to combine robot movements with emotional expression in HRI is possible [40,41].
To assess the user’s perception of the emotional body movements of a robot, subjective and objective evaluations can be considered. Subjective evaluation is the main approach for evaluating the bodily emotional expressions of robots, generally adopting self-reported methods such as questionnaires, interviews, and so on. Users are asked to watch the body movements of the robot, then choose and rate the emotions they feel according to pre-determined categories. In the studies of Tsiourti et al. [28], Erden [29], Fu et al. [36], Wei et al. [40], Li et al. [41], Urakami [43], and others, different questionnaires were used to evaluate the perceptions of robot emotional body movements from multiple views, such as emotion understandability, anthropomorphism, likeability, and so on. Objective evaluations involve obtaining user emotional responses through physiological measures, such as eye-tracking, electroencephalogram (EEG), electrocardiogram (ECG), electrodermal activity (EDA), and so on [44,45]. Related studies focused on objective evaluation in HRI have been increasing in recent years. For example, Guo et al. studied the influence of a humanoid robot’s emotional behaviors on the emotional responses of users through subjective reporting, eye-tracking, and EEG [46]. Staffa et al. collected the EEG signals of 10 participants demonstrating either a positive or negative personality who interacted with a Pepper robot, and demonstrated the possibility of classifying a user’s emotional responses to the robot’s behavior from the EEG signals [47].

2.2. The Neural Network Used for Emotion Recognition in HRI

The aim of the computational model to be built in this study is to construct a mapping between the robot’s emotional body movements and corresponding emotional states. The use of a neural network is a feasible approach for this purpose, due to their non-linear, self-organization, and self-adaptation characteristics, among others. Neural networks are often used for emotion recognition in different contexts [48,49,50]. The design for emotional body movements of robots and the corresponding perceptions of user emotions can be used as the inputs and outputs, respectively, for the computational model to be developed. To date, multiple neural networks have been adopted for emotion recognition via facial expression, body gesture, and speech in the HRI context [51,52,53,54], but those focused on emotion recognition with respect to robot body movements remain rare. In the work of Li et al., an artificial emotion expression method that could recognize the emotion patterns expressed by a robot according to its motion behaviors was proposed, for which the generalized regression neural network (GRNN) and the discrete octagonal Pleasure–Arousal–Dominance (PAD) model were used [41]. Furthermore, some studies in the field of facial expression recognition in HRI have focused on the relationships between facial actions and emotions, which can serve as a reference, to some extent. For example, Rawal et al. introduced the expression generation network (ExGenNets), which can reconstruct simplified facial images from robot joint configurations with a classifier network for facial expression recognition [55]. Khashman proposed an emotional back propagation (BP) neural network with two emotional parameters—anxiety and confidence—for facial recognition [56]. Li et al. proposed an emotion recognition system for a humanoid robot based on a convolutional neural network (CNN) and six basic emotions—happiness, anger, disgust, fear, sadness, and surprise—and used a long short-term memory (LSTM) recurrent neural network to build the relationship between the transformation of facial expressions and the six basic emotions [57]. Li et al. built a facial expression and touch gesture emotion dataset and proposed a multi-eigenspace-based multimodal fusion network (MMFN) to recognition tactile–visual bimodal emotion in the HRI context [58]. The results of these studies demonstrate that neural networks can be used to construct the relationship between robot motions and human emotion, providing a feasible approach to develop the corresponding computational model.

3. Design of Robot’s Emotional Body Movements

3.1. Materials

A humanoid robot named Yanshee was employed to imitate the emotional body movements of human. The Yanshee robot utilizes the open hardware platform architecture of Raspberry Pi + STM32, and consists of 17 DoFs. The body movements that the Yanshee robot imitates come from two databases: a human body kinematic dataset constructed by Zhang et al. and a database for the bodily expressive action stimulus test (BEAST) constructed by De Gelder et al. In the work of Zhang et al., 22 semi-professional actors provided a total of 1402 recordings of body movements, with happiness, sadness, anger, fear, disgust, surprise, and neutral emotional expressions [26]. The database constructed by De Gelder et al. is composed of 254 whole-body expressions from 46 actors expressing emotions, including happiness, sadness, anger, and fear [32]. A total of 72 body movements reflecting 6 emotions—including happiness, fear, anger, disgust, sadness, and surprise—are mimicked by Yanshee (i.e., 12 for each emotion), with 32 of the body movements being imitated from the work of Zhang et al., and 40 of them from the work of De Gelder et al. The videos of these body movements were labelled from V1 to V72. Examples of the robot’s body movements imitating different emotions are shown in Figure 2. In Figure 2a–f, the videos V10, V8, V4, V12, V2, and V6 correspond to the pre-determined emotions of happiness, fear, anger, disgust, sadness, and surprise, respectively.
The behaviors of the humans and robots in Figure 2 visually illustrate the imitation of human emotional behaviors by the Yanshee robot. When imitating, some difficulties were found due to the limited DOFs of Yanshee. For example, the contrast between human and Yanshee robot poses are quite clear in Figure 2d. The Yanshee robot cannot achieve the sloping shoulder posture, as it does not possess the associated activity structures. Although it makes little difference while posing, the emotional perception of the user may change [59].

3.2. Subjective Evaluation for User’s Perception of Emotion

Considering that robots are limited by their physical structure, imitating human body movements does not necessarily make users feel the same emotion as that in the datasets. Hence, we need to verify whether the emotions identified belong to the pre-determined emotional categories. A total of 51 students with age ranging from 20 to 27 years (M = 23.55, SD = 1.39) were recruited from the same university, including 26 males and 24 females. All of them had normal or corrected-to-normal vision and were self-reported to be free from any history of neurological or psychiatric disorders. They participated after being fully informed of the experimental procedures and were paid. They were asked to watch 72 videos, with each video lasting about 2 s. After each video, they were asked to score the emotion reflected by the robot’s body movement (out of happiness, fear, anger, disgust, sadness, or surprise), followed by a short break. The order of the six emotional categories was random. A 7-point Likert scale was used, where 7 means strong and 1 means weak. The experiments were implemented in a well-lit and distraction-free room. Each participant spent an average of about 40 min. A scenario in which a student was watching video V41 is shown in Figure 3. All data obtained from the subjective evaluations were anonymized.

3.3. Results and Analysis

Each bodily expression of the robot shown in the video was scored under different emotional categories, where the scores indicate the intensity of emotion expressed by the robot’s body movement from the user’s view. The data were basically consistent with a normal distribution. The repeated-measures analysis of variance (ANOVA) was conducted to analyze these data. The data were divided into six groups, according to the scores in different emotional categories for all robot body movements. The Mauchly’s test of sphericity p < 0.05. In the tests for within-subject effects, F (5, 300) = 13.062, p < 0.001 partial η2 = 0.179. These results reveal that there exist significant differences in different emotional categories. In the pairwise comparisons, the scores for surprise showed a significant difference with the scores for other emotions, and the scores for happiness showed a significant difference from the scores for sadness. The mean values (M values) of scores for all robot body movements and for each emotion category are shown in Figure 4a. The overall trend shows relatively high scores for surprise and relatively low scores for sadness, while the scores were similar for happiness, fear, anger, and disgust. This indicates that, for all the designed robot body movements, expressions of surprise are easier to perceive than others, while expressions of sadness are harder to perceive than others.
The emotion recognition rate obtained in the subjective evaluation is defined as the ratio of the number of participants who score the highest for this emotion category to the total number of participants, and the data are shown in Figure 4b. The recognition rates for happiness were relatively better, but the overall trends were not high, especially for sadness, fear, and disgust. The Spearman correlation test was used to analyze the correlations between the recognition rates and the scores of robot body movements, and the results are shown in Table 1, where s represents the Spearman correlation coefficient. All p-values are smaller than 0.001. Happiness presented the strongest correlation, with s = 0.97. However, for surprise, the correlation value was only s = 0.569. Although there were relatively large differences in the correlations for different emotions, the results reveal that there are positive correlations with different intensities between the recognition rates and the scores of robot body movements in the six emotional categories. For example, the recognition of happiness was highly influenced by the intensity of emotion with happiness, while that of surprise was relatively less influenced by the intensity of surprise. For the robot’s emotional body movements, the impact of the intensities of emotions on the recognitions of negative emotions (including disgust, fear, sadness, and anger) were less than that for the recognition of happiness.
Hence, to help the users to recognize the emotions reflected by robot body movements easier, we built a dataset of robot emotional body movements by taking the recognition rate of more than 50% and the score of more than 3.5 as conditions. The robot body movements included in the dataset and the corresponding emotion recognition rate, M value, standard deviation (SD), etc., are detailed in Table 2. Then, 25 robot emotional body movements were selected, which better reflect the association between the robot’s body movements and the corresponding emotions. It should be noted that the eligible robot’s body movements after screening belonged to four emotional categories: happiness, anger, sadness, and surprise. We observed some differences between the initial design and the results. Some body movements of the Yanshee robot were designed to express a certain emotional state but were identified as another emotional state. For example, the body movement shown in V4 was designed to represent anger but was identified as happiness with a high recognition rate (86.27%). The body movement shown in V8 was designed for fear, but was also identified as happiness with a recognition rate of 74.51%. The body movement shown in V10 was designed for happiness but was identified as fear with a recognition rate of 37.25%.

4. Emotion Recognition Model Using BP Neural Network for Robot’s Body Movement

4.1. The Neural Network Topological Architecture of the Proposed Model

The emotional recognition of robot body movements can be regarded as the mapping from the input layer to the output layer in a neural network, where the body movements of the robot and corresponding emotions can be used as the inputs and outputs of the neural network in the proposed model, respectively. In training the neural network, the proposed model gradually improves its emotion recognition ability through learning the mapping between characteristics of the robot’s body movements and the corresponding emotions. The architecture of the proposed model is given by Figure 5. The input layer consists of the information expressed by the robot’s body movements, including the moving body part, the direction of movement, and the moving mode. The output layer shows the emotions including happiness, fear, anger, disgust, sadness and surprise.

4.2. Encoding Rules

The robot emotional body movements in the dataset built in Section 3.3 had to be encoded to transform the movements into numerical forms, allowing them to be applied as the inputs in neural network training. When the body part, the direction of movement, or the moving mode is triggered, the corresponding node in the input layer of the neural network is set to 1, and the other nodes (being untriggered) are set to 0. Similarly, the node corresponding to the triggered emotion in the output layer is set to 1, while the other emotion nodes are set to 0. The motion state set for inputs and the emotion state set for outputs are illustrated by Table 3. The motion state set can be further divided into three subsets, reflecting the body part, moving direction, and moving mode. Xn represents the input information in the motion state set, where n is the sequence number of the robot’s body movements. Xn = {On, Dn, Mn}, where O, D, and M represent the body part, moving direction, and moving mode sets, respectively. The body parts of the Yanshee robot are shown in Figure 6. Yn = {y1, y2, y3, y4, y5, y6} represents the information in the emotion state set, where y1, y2, y3, y4, y5, and y6 represent happiness, fear, anger, disgust, sadness, and surprise, respectively. n [ 1 ,   72 ] is a positive integer. Xn is the input coded sequence, which is used in the input layer of the neural network shown in Figure 5, while Yn is the corresponding output coded sequence.
In the studies of Fu et al. [31], Yagi et al. [60], and Mahzoon et al. [61], the vertical features of robot body movements were considered to express emotions, and the studies of de Meijer [33] and De Silva et al. [34] referred to both vertical and lateral features, including the stretching and bowing movements of the body, to distinguish positive and negative emotions. Hence, when encoding the robot’s emotional body movements, the vertical and lateral movements are considered separately. Take the robot’s body movement in V4 as an example. In Figure 7, the body movements of the Yanshee robot over two seconds can be seen, which consist of two movements combined. One is from the vertical view, where the Yanshee robot’s head, torso, and left foot rotate to the front-left. The other is from the lateral view, where the robot’s hands are stretched forward simultaneously.
As the triggered and untriggered nodes are set to 1 and 0, respectively, considering Table 3, for the body parts, the nodes of the robot’s head, torso, left foot, left hand, and right hand are set to 1, while the node of the right foot is set to 0. For the moving direction, left and forward are 1 and the others are 0, and for the moving mode, rotation and extend are 1 and the others are 0. P4, D4, and M4 are given by Equation (1). X4 = {P4, D4, M4}.
P 4 = 1 , 1 , 0 , 1 , 1 , 1 } D 4 = 1 , 0 , 1 , 0 , 0 , 0 , 0 , 0 } M 4 = 1 , 0 , 0 , 1 , 0 , 0 }
After all the movements are encoded, the corresponding emotional states are also considered. For example, for the robot’s body movement in V4, the corresponding emotion category is happiness, which means that Y4 = {1, 0, 0, 0, 0, 0}.

4.3. The Proposed Emotional Body Movement Recognition Model Using BP Neural Network

As the robot’s emotional body movements designed in Section 3 are insufficient for recognizing the robot’s body movements with respect to the six emotion categories, we introduce our previous design reflecting twelve emotional body movements of the Yanshee robot, which has been verified to express the emotions of happiness, anger, sadness, fear, and disgust. In total, 37 emotional body movements of the Yanshee robot were employed, of which 28 were used to train the proposed neural network and 9 were used for the recognition experiments. The number of neurons in the hidden layer was determined using the Levenberg–Marquardt method, and the trend of the corresponding value of mean square error (MSE) is shown in Figure 8. Considering the minimum value of MSE, the corresponding number of neurons in the hidden layer was set to 13. Hence, the proposed emotion recognition model uses a BP neural network including an input layer with 20 nodes, a hidden layer with 13 nodes, and an output layer with 6 nodes. For convenience, the proposed emotion recognition model using a BP neural network is abbreviated as the EBMR-BP model.
The activation function, logsig(x), was used, as it can map the inputs into the range of (0, 1) to match the encoding for the robot’s body movements. The activations a and net activations z are given by Equations (2) and (3), respectively.
a ( l ) = l o g s i g W ( l ) a ( l 1 ) + b ( l ) ,
z ( l ) = W ( l ) l o g s i g z ( l 1 ) + b ( l ) ,
where l = 1, 2, which represents the number of neural network layers; a ( 0 ) = X ; and a ( 2 ) is the output of the neural network. The error of the lth neural layer, δ ( l ) , is shown in Equation (4):
δ ( l ) = l o g s i g · z ( l ) b ( l ) δ ( l + 1 )
W and b are the weight and bias matrices, respectively, which are updated using the LM method. The learning rate is set to 0.01, and the goal MSE is set to 0.00001. A total of 28 robot movements were randomly selected to train the proposed BP neural network, and some examples of the corresponding encoded inputs and outputs are illustrated in Table 4. The percentages of the training, validation, and test sets were 80%, 10%, and 10%, respectively. After training the EBMR-BP model, the MSE trend is shown in Figure 9. The minimum value of MSE was 0.219, which is not small enough.

5. The Optimization for Proposed Model Using PSO and RMSProp

The particle swarm optimization (PSO) method and the root mean square propagation (RMSProp) algorithm are used to optimize the proposed EBMR-BP model. The main aim of using PSO and RMSProp is to reduce the values of MSEs. The key steps are summarized as follows.
Step 1. Initialize the weights and biases of the proposed BP neural network, as well as the velocity and the position of particles to build the PSO-BP neural network. The positions of particles represent the weights and biases of the neural network, and the speed controls the corresponding state of movement in the search space. In the space of particles, the position matrix L and the speed matrix v of a particle are calculated using Equations (5) and (6), respectively.
v s t + 1 = ω v s t + c 1 r 1 p s t L s t + c 2 r 2 p g t L s t ,
L s t + 1 = L s t + v s t + 1 ,
where s is the serial number of particles, t is the count of iterations, ps is the personal best (pBest), and pg is the global best (gBest). The learning factors c1 and c2 are set as 1.6 and 2, respectively; the random factors r1 and r2 are distributed between 0 and 1; and the inertia weight ω is set to be 0.6.
Step 2. Compute the fitness for particles in the space. The MSEs between the outputs of the proposed BP neural network and the target values are used as the fitness function. The speed of particles is limited by Equation (7), in order to ensure their convergence.
v s t = m a x m i n v s t , v m a x , v m i n ,
where vmax and vmin are set to 0.5 and −0.5, respectively. Then, pBest and gBest are updated, and the position L and the speed v can be updated by Equations (2) and (3) until convergence or the maximum number of iterations is reached.
Step 3. The errors of each layer in the neural network are calculated, and the RMSProp algorithm is introduced when training the neural network. For the l-th layer, the gradients for W and b are given by Equations (8) and (9), respectively.
L W ( l ) = δ l ( a ( l 1 ) ) T ,
L b ( l ) = δ l ,
where L represents the loss function. For the t-th iteration, the accumulations of gradient squares are given by Equation (10).
G t = β G t 1 + ( 1 β ) g t · g t ,
where the attenuation rate β = 0.9 and g t represents the gradients of parameters. Then, W and b can be updated until the stop condition is reached.
The above-mentioned process consists of training the PSO-BP neural network, then carrying out optimization using the RMSProp algorithm. For simplicity, it is abbreviated as the PSO-BP-RMSProp model. The main flow of this model is shown in Figure 10.

6. Experiments and Comparisons

Considering the EBMR-BP model proposed in Section 4.3 and steps 1 and 2 in Section 5, it is a process that involves optimizing the EBMR-BP model using the PSO method. When just considering the EBMR-BP model and step 3 in Section 5, the EBMR-BP model is optimized using the RMSProp algorithm. To simplify for discussion, these two are shortened to the PSO-BP model and BP-RMSProp model, respectively. Four models, including the EBMR-BP model, the PSO-BP model, the BP-RMSProp model, and the PSO-BP-RMSProp model, were used to recognize the emotions associated with the robot’s body movements. Nine robot emotional body movements were employed as the materials for recognition. The maximum number of iterations was 200 for the PSO-BP and PSO-BP-RMSProp models. For the BP-RMSProp model, the iterations stopped in the case of reaching the target error. The trends of MSEs for the PSO-BP, BP-RMSProp, and PSO-BP-RMSProp models with the epochs are shown in Figure 11a, Figure 11b, and Figure 11c, respectively. As shown in Figure 11b, the convergence ability of the BP-RMSProp model is not as good. The minimum MSEs for the PSO-BP, BP-RMSProp, and PSO-BP-RMSProp models were 0.0664, 9.9576 × 10−6, and 0.0664, respectively, and it was observed that the MSEs were smaller than that in Figure 9 in the case of convergence. This means that PSO and RMSProp are useful for reducing the MSE. The heatmaps of the data obtained with the EBMR-BP, PSO-BP, BP-RMSProp, and PSO-BP-RMSProp models are illustrated in Figure 12, where the horizontal axis represents different categories of emotion described in Section 4.2, and the vertical axis represents the serial number of the robot’s different body movements. The horizontal position of the maximum value for each row shows the corresponding emotional state of the robot’s body movement. It is obvious that, for the EBMR-BP model, the differences in the data were quite small, as shown in Figure 12a.
A convolutional neural network (CNN) was used to compare the emotion recognition performance for the robot’s body movements. The architecture has a total of 12 layers. The videos used in Section 4.3 were sampled into images, where the sampling frequency was 30 kHz, and the input layer of the CNN takes these images as input. Then, a 2D convolutional layer with 16 filters and a kernel size of 3 × 3, a batch normalization layer, a ReLU layer, and a max pooling layer follow. The next four layers include another 2D convolutional layer with 32 filters and a kernel size of 3 × 3, a batch normalization layer, a ReLU layer, and a max pooling layer. These are followed by a dense layer with six nodes, which represents the six emotional categories shown in Section 4.2. The following Softmax activation layer and classification layer yield the output of the CNN. The accuracy and loss for each epoch of the CNN are shown in Figure 13a and Figure 13b, respectively. The comparisons between different methods for emotion recognition from robot body movements are given in Table 5, where Method I, II, and III represent the subjective emotion evaluation, emotion recognition using the proposed models, and emotion recognition using the CNN, respectively.
The subjective emotion recognition results were taken as the baseline. The numbers that could be recognized by the four proposed models were 6, 5, 4, and 8, respectively, while that of the CNN was 6. The emotion recognition rates of the EBMR-BP, PSO-BP, BP-RMSProp, PSO-BP-RMSProp, and CNN models were 66.67%, 55.56%, 44.44%, 88.89%, and 66.67%, respectively. This means that the proposed EBMR-BP model is effective and, when the PSO method and RMSProp algorithm are introduced separately to optimize the proposed BP model, they do not yield appropriate performances. However, when they are used in combination, the MSEs generated by the PSO-BP-RMSProp model are smaller than that generated by the EBMR-BP model, the convergence ability is the same as the PSO-BP model, and the corresponding emotion recognition rates are also improved. This indicates that optimization of the proposed model using PSO and RMSProp significantly improved its emotion recognition performance. The CNN method can be used for recognition, and its emotion recognition rate was the same as that of the EBMR-BP model. However, considering the results shown in Table 5, it is uncertain whether the number of robot body movements with other emotions (except for happiness) in the training dataset are sufficient for the CNN method. Extension of the dataset may lead to a better recognition rate for CNN, but the proposed PSO-BP-RMSProp model was able to achieve higher recognition rates with only the presently used data.

7. Conclusions

This study mainly focused on an emotional recognition method for a humanoid robot’s body movements. The Yanshee robot was employed, and its emotional body movements were designed by imitating the human body movements shown in the works of Zhang et al. and De Gelder et al. Subjective questionnaires were used, and the corresponding data were analyzed using statistical methods, including the repeated ANOVA and the Spearman correlation methods, in order to explore user perceptions and select appropriate robot emotional body movements. Then, a neural network topological architecture was built. The selected movements and corresponding emotions were used and encoded as its inputs and outputs, respectively. In this manner, an emotional body movement recognition model using a BP neural network for robot body movements—called the EBMR-BP model for convenience—was proposed. To optimize the proposed model, the PSO method and the RMSProp algorithm were introduced to reduce the MSE and improve the emotional body movement recognition rates. The PSO-BP-RMSProp model was developed, and the corresponding key steps were illustrated. The implementations of PSO-BP and BP-RMSProp were also induced on this basis. Experiments and comparisons of the emotional recognition of robot body movements were conducted using the EBMR-BP, PSO-BP, BP-RMSProp, and PSO-BP-RMSProp models, as well as a CNN model for comparison. The main contributions of this study are summarized as follows:
  • Positive correlations between the emotion recognition rates of robot body movements and the intensities of their emotional expressions were found. Moreover, the impacts on the emotion recognition rates differed for different emotions expressed by the robot’s body movements. The happiness recognition rates of robot body movements were greatly affected by the happiness expressions. Meanwhile, the influence of the robot’s negative emotion-related expressions on negative emotion recognition rates was lower than that of the happiness recognition expressions on the happiness recognition rates. Compared with other emotions, the expression of surprise showed the smallest relative effect on the recognition rates of surprise.
  • A dataset consisting of 25 robot emotional body movements and corresponding emotional states was designed, and it can be used as research material in future work focused on HRI.
  • The EBMR-BP model, which provides a feasible and effective approach to recognizing the emotions associated with a humanoid robot’s body movements, was proposed. The topological architecture, the encoding rule, and other aspects were described in detail. It allows for a mapping between a robot’s emotional body movements and human emotions to be successfully built using a computational model.
  • Optimization of the EBMR-BP model using a combination of the PSO method and the RMSProp algorithm was proposed to illustrate the feasible optimization path and verify the improvement in the emotion recognition ability of the robot’s emotional body movements. It also provides a feasible approach to achieve a relatively higher emotion recognition rate associated with body movements (88.89%) with a relatively small amount of data.
Overall, the findings of this study can provide beneficial references for future studies on self-adapting emotional recognition for robot body movements, as well as for the design of emotional interactions in the HRI context. Limitations are described from five aspects. First, the subjective results of robot body movements were obtained according to data obtained from young people. If the participants were from a wider range, the findings of this study may be more widely generalizable. Second, the participants of the subjective evaluation were Chinese. It is not certain whether the results would remain consistent in a different country or culture. Third, some subtle expressions of human body movements are hard to imitate when using the Yanshee robot due to its limited physical structure. This led to relatively large differences between the pre-designed robot body movements and the user perception results detailed in Section 3.2. Although appropriate body movements of the robot were selected in this work, more useable data may be obtained when using a robot with a higher number of DoFs. However, attaining such technical conditions was not possible in this study. Fourth, the effect of the moving speed of the robot’s joints on the subjective emotion evaluation was not considered, as it is uncertain how the variation in movement speed in different parts of robots could affect the evaluation. However, the speed may have some impacts, as some studies have shown that variation in robot velocity when moving forwards has impacts on human understanding [62], the positive and negative perceptions of observers [63], etc. Fifth, the Yanshee robot is a biped robot. As such, the encoding rules of the EBMR-BP and PSO-BP-RMSProp models in this study are in line with the characteristics of biped robots. In our view, their portability to other robot platforms is possible in theory. One reason is that our design was proposed through the imitation of human emotional body movements. Another reason is that robot body movements reflecting the same emotion show similar characteristics [33,34]. For a wheeled humanoid robot or other robot platforms with different physical characteristics, corresponding empirical studies are required. Our future work will focus on expanding the used dataset in order to develop models with better robustness and generalization ability, as well as further improving the emotion recognition rate of body movements.

Author Contributions

Conceptualization, formal analysis, methodology, writing—draft preparation, W.G.; investigation, formal analysis, visualization, T.J.; visualization, W.Z.; project administration, F.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 52105262, U2013602.

Institutional Review Board Statement

Ethical review and approval were waived for this study, due to the fact that the experiments in the study were not invasive examinations, did not contradict any ethical standards, and did not ask for any private information.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Acknowledgments

The authors would like to thank all of the participants that took the time and effort to participate in testing.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Błażejowska, G.; Gruba, Ł.; Indurkhya, B.; Gunia, A. A Study on the Role of Affective Feedback in Robot-Assisted Learning. Sensors 2023, 23, 1181. [Google Scholar] [CrossRef] [PubMed]
  2. Han, H.; Karadeniz, O.; Dalyan, T.; Sonmez, E.B.; Sarioglu, B. Acknowledge of Emotions for Improving Student-Robot Interaction. Intell. Autom. Soft Comput. 2023, 37, 1209–1224. [Google Scholar] [CrossRef]
  3. James, J.; Balamurali, B.T.; Watson, C.I.; MacDonald, B. Empathetic Speech Synthesis and Testing for Healthcare Robots. Int. J. Soc. Robot. 2021, 13, 2119–2137. [Google Scholar] [CrossRef]
  4. Spekman, M.L.C.; Konijn, E.A.; Hoorn, J.F. How Physical Presence Overrides Emotional (Coping) Effects in HRI: Testing the Transfer of Emotions and Emotional Coping in Interaction with a Humanoid Social Robot. Int. J. Soc. Robot. 2021, 13, 407–428. [Google Scholar] [CrossRef]
  5. Lee, M.; Lee, J.H.; Lee, M. RU-FEMOIN—A Database of Facial Expressions: Utilization for Robotic Mood Transition of Patient Robots. Expert Syst. Appl. 2023, 234, 120907. [Google Scholar] [CrossRef]
  6. Chen, X.; Sun, S.; Zhang, Z.; Ma, Z.; Wu, X.; Li, H.; Chen, T.; Zhang, K. Consumer Shopping Emotion and Interest Database: A Unique Database with a Multimodal Emotion Recognition Method for Retail Service Robots to Infer Consumer Shopping Intentions Better Than Humans. J. Electron. Imaging 2022, 31, 061807. [Google Scholar] [CrossRef]
  7. Yoo, S.R.; Kim, S.H.; Jeon, H.M. How Does Experiential Value Toward Robot Barista Service Affect Emotions, Storytelling, and Behavioral Intention in the Context of COVID-19? Sustainability 2022, 14, 450. [Google Scholar] [CrossRef]
  8. Yang, K.; Wang, H. The Application of Interactive Humanoid Robots in the History Education of Museums Under Artificial Intelligence. Int. J. Hum. Robot. 2023, 20, 2250016. [Google Scholar] [CrossRef]
  9. Cameron, D.; Millings, A.; Fernando, S.; Collins, E.C.; Moore, R.; Sharkey, A.; Evers, V.; Prescott, T. The Effects of Robot Facial Emotional Expressions and Gender on Child–Robot Interaction in a Field Study. Connect. Sci. 2018, 30, 343–361. [Google Scholar] [CrossRef]
  10. Hashimoto, M.; Kondo, H.; Tamatsu, Y. Gaze Guidance Using a Facial Expression Robot. Adv. Robot. 2009, 23, 1831–1848. [Google Scholar] [CrossRef]
  11. Val-Calvo, M.; Álvarez-Sánchez, J.R.; Ferrández-Vicente, J.M.; Fernandez, E. Affective Robot Story-Telling Human-Robot Interaction: Exploratory Real-Time Emotion Estimation Analysis Using Facial Expressions and Physiological Signals. IEEE Access 2020, 8, 134051–134066. [Google Scholar] [CrossRef]
  12. Filippini, C.; Perpetuini, D.; Cardone, D.; Merla, A. Improving Human–Robot Interaction by Enhancing Nao Robot Awareness of Human Facial Expression. Sensors 2021, 21, 6438. [Google Scholar] [CrossRef] [PubMed]
  13. Park, J.W.; Lee, H.S.; Chung, M.J. Generation of Realistic Robot Facial Expressions for Human Robot Interaction. J. Intell. Robot. Syst. 2015, 78, 443–462. [Google Scholar] [CrossRef]
  14. Oh, K.; Jang, M.; Kim, S.J. Automatic Emotional Expression of a Face Robot by Using a Reactive Behavior Decision Model. J. Mech. Sci. Technol. 2010, 24, 769–774. [Google Scholar] [CrossRef]
  15. Dou, X.; Wu, C.F.; Lin, K.C.; Gan, S.; Tseng, T.-M. Effects of Different Types of Social Robot Voices on Affective Evaluations in Different Application Fields. Int. J. Soc. Robot. 2021, 13, 615–628. [Google Scholar] [CrossRef]
  16. Lee, J. Generating Robotic Speech Prosody for Human Robot Interaction: A Preliminary Study. Appl. Sci. 2021, 11, 3468. [Google Scholar] [CrossRef]
  17. Sawabe, T.; Honda, S.; Sato, W.; Ishikura, T.; Kanbara, M.; Yoshikawa, S.; Fujimoto, Y.; Kato, H. Robot Touch with Speech Boosts Positive Emotions. Sci. Rep. 2022, 12, 6884. [Google Scholar] [CrossRef]
  18. Fu, C.; Deng, Q.; Shen, J.; Mahzoon, H.; Ishiguro, H. A Preliminary Study on Realizing Human–Robot Mental Comforting Dialogue Via Sharing Experience Emotionally. Sensors 2022, 22, 991. [Google Scholar] [CrossRef]
  19. Homma, T.; Sun, Q.; Fujioka, T.; Takawaki, R.; Ankyu, E.; Nagamatsu, K.; Sugawara, D.; Harada, E.T. Emotional Speech Synthesis for Companion Robot to Imitate Professional Caregiver Speech. arXiv 2021, arXiv:2109.12787. [Google Scholar]
  20. Kuo, Y.C.; Tsai, P.H. Enhancing Expressiveness of Synthesized Speech in Human-Robot Interaction: An Exploration of Voice Conversion-Based Methods. In Proceedings of the 2024 10th International Conference on Control, Automation and Robotics (ICCAR), Singapore, 27–29 April 2024. [Google Scholar]
  21. Chang, Y.; Ren, Z.; Zhang, Z.; Jing, X.; Qian, K.; Shao, X.; Hu, B.; Schultz, T.; Schuller, B.W. STAA-Net: A Sparse and Transferable Adversarial Attack for Speech Emotion Recognition. arXiv 2024, arXiv:2402.01227. [Google Scholar] [CrossRef]
  22. Zeng, Z.; Pantic, M.; Roisman, G.I.; Huang, T.S. A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 39–58. [Google Scholar] [CrossRef] [PubMed]
  23. Torre, I.; Holk, S.; Yadollahi, E.; Leite, I.; McDonnell, R.; Harte, N. Smiling in the Face and Voice of Avatars and Robots: Evidence for a ‘smiling McGurk Effect’. IEEE Trans. Affect. Comput. 2022, 15, 393–404. [Google Scholar] [CrossRef]
  24. Striepe, H.; Donnermann, M.; Lein, M.; Lugrin, B. Modeling and Evaluating Emotion, Contextual Head Movement and Voices for a Social Robot Storyteller. Int. J. Soc. Robot. 2021, 13, 441–457. [Google Scholar] [CrossRef]
  25. McColl, D.; Nejat, G. Recognizing Emotional Body Language Displayed by a Human-Like Social Robot. Int. J. Soc. Robot. 2014, 6, 261–280. [Google Scholar] [CrossRef]
  26. Zhang, M.; Yu, L.; Zhang, K.; Du, B.; Zhan, B.; Chen, S.; Jiang, X.; Guo, S.; Zhao, J.; Wang, Y.; et al. Kinematic Dataset of Actors Expressing Emotions. Sci. Data 2020, 7, 292. [Google Scholar] [CrossRef] [PubMed]
  27. De Gelder, B. Why Bodies? Twelve Reasons for Including Bodily Expressions in Affective Neuroscience. Philos. Trans. R. Soc. B 2009, 364, 3475–3484. [Google Scholar] [CrossRef]
  28. Tsiourti, C.; Weiss, A.; Wac, K.; Vincze, M. Multimodal Integration of Emotional Signals from Voice, Body, and Context: Effects of (in) Congruence on Rmotion Tecognition and Attitudes Towards Robots. Int. J. Soc. Robot. 2019, 11, 555–573. [Google Scholar] [CrossRef]
  29. Erden, M.S. Emotional Postures for the Humanoid-Robot Nao. Int. J. Soc. Robot. 2013, 5, 441–456. [Google Scholar] [CrossRef]
  30. Takahashi, Y.; Kayukawa, Y.; Terada, K.; Inoue, H. Emotional Expressions of Real Humanoid Robots and Their Influence on Human Decision-Making in a Finite Iterated Prisoner’s Dilemma Game. Int. J. Soc. Robot. 2021, 13, 1777–1786. [Google Scholar] [CrossRef]
  31. Fu, C.; Wang, S.; Li, Z.; Gupta, A.; Meneses, A.; Ishiguro, H.; Yoshikawa, Y. Modulating Perceived Authority and Warmth of Mobile Social Robots through Bodily Openness and Vertical Movement in Gait. IEEE Robot. Autom. Let. 2024, 9, 7971–7978. [Google Scholar] [CrossRef]
  32. De Gelder, B.; Van den Stock, J. The Bodily Expressive Action Stimulus Test (BEAST). Construction and Validation of a Stimulus Basis for Measuring Perception of Whole Body Expression of Emotions. Front. Psychol. 2011, 2, 181. [Google Scholar] [CrossRef] [PubMed]
  33. De Meijer, M. The Contribution of General Features of Body Movement to the Attribution of Emotions. J. Nonverbal Behav. 1989, 13, 247–268. [Google Scholar] [CrossRef]
  34. De Silva, P.R.; Bianchi-Berthouze, N. Modeling Human Affective Postures: An iInformation Theoretic Characterization of Posture Features. Comput. Animat. Virtual Worlds 2004, 15, 269–276. [Google Scholar] [CrossRef]
  35. Halovic, S.; Kroos, C. Not All Is Noticed: Kinematic Cues of Emotion-Specific Gait. Hum. Mov. Sci. 2018, 57, 478–488. [Google Scholar] [CrossRef] [PubMed]
  36. Fu, C.; Alexis, M.; Yoshikawa, Y.; Ishiguro, H. Enhancing the Mobile Humanoid Robot’s Emotional Expression with Affective Vertical-Oscillations. Int. J. Soc. Robot. 2024, 16, 1523–1540. [Google Scholar] [CrossRef]
  37. Häring, M.; Bee, N.; André, E. Creation and Evaluation of Emotion Expression with Body Movement, Sound and Eye Color for Humanoid Robots. In Proceedings of the IEEE International Symposium on Robot & Human Interactive Communication, Venice, Italy, 6–8 September 2011. [Google Scholar]
  38. Tuyen, N.T.V.; Elibol, A.; Chong, N.Y. Learning Bodily Expression of Emotion for Social Robots Through Human Interaction. IEEE Trans. Cogn. Dev. Syst. 2020, 13, 16–30. [Google Scholar] [CrossRef]
  39. Coulson, M. Attributing Emotion to Static Body Postures: Recognition Accuracy, Confusions, and Viewpoint Dependence. J. Nonverbal Behav. 2004, 28, 117–139. [Google Scholar] [CrossRef]
  40. Wei, Y.; Zhao, J. Designing Robot Behavior in Human Robot Interaction Based on Emotion Expression. Ind. Robot. Int. J. 2016, 43, 380–389. [Google Scholar] [CrossRef]
  41. Li, L.; Zhao, Z. Designing Behaviors of Robots Based on the Artificial Emotion Expression Method in Human–Robot Interactions. Machines 2023, 11, 533. [Google Scholar] [CrossRef]
  42. Gong, S.; Zhao, J.; Zhang, Z.; Xie, B. Task Motion Planning for Anthropomorphic Arms Based on Human Arm Movement Primitives. Ind. Robot. 2020, 47, 669–681. [Google Scholar] [CrossRef]
  43. Urakami, J. Do Emotional Robots Get More Help? How a Robots Emotions Affect Collaborators Willingness to Help. Int. J. Soc. Robot. 2023, 15, 1457–1471. [Google Scholar] [CrossRef]
  44. Hu, F.; Zhang, L.; Yang, X.; Zhang, W.-A. EEG-Based Driver Fatigue Detection Using Spatio-Temporal Fusion Network with Brain Region Partitioning Strategy. IEEE Trans. Intell. Transp. 2024, 25, 9618–9630. [Google Scholar] [CrossRef]
  45. Si, X.; Huang, H.; Yu, J.; Ming, D. The fNIRS-based Emotion Recognition by Spatial Transformer and WGAN Data Augmentation Towards Developing A Novel Affective BCI. IEEE Trans. Affect. Comput. 2024, 14, 1–15. [Google Scholar] [CrossRef]
  46. Guo, F.; Li, M.M.; Qu, Q.X. The Effect of a Humanoid Robot’s Emotional Behaviors on Users’ Emotional Responses: Evidence from Pupillometry and Electroencephalography Measures. Int. J. Hum.-Comput. Int. 2019, 35, 1947–1959. [Google Scholar] [CrossRef]
  47. Staffa, M.; D’Errico, L.; Sansalone, S.; Alimardani, M. Classifying Human Emotions in HRI: Applying Global Optimization Model to EEG Brain Signals. Front. Neurorobotics 2023, 17, 1191127. [Google Scholar] [CrossRef]
  48. Zhang, S.; Zhang, J.; Song, W.; Yang, L.; Zhao, X. Hierarchical-Attention-Based Neural Network for Gait Emotion Recognition. Phys. A Stat. Mech. Its Appl. 2024, 637, 129600. [Google Scholar] [CrossRef]
  49. Wang, Q.; Wang, J.; Zhu, M. Research on Emotion Recognition Algorithm Based on Improved BP Neural Network. In Proceedings of the 27th ECS Meeting, Montréal, QC, Canada, 18–22 May 2025. [Google Scholar]
  50. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed]
  51. Cai, Y.; Li, X.; Li, J. Emotion Recognition Using Different Sensors, Emotion Models, Methods and Datasets: A Comprehensive Review. Sensors 2023, 23, 2455. [Google Scholar] [CrossRef]
  52. Dwijayanti, S.; Iqbal, M.; Suprapto, B.Y. Real-Time Implementation of Face Recognition and Emotion Recognition in a Humanoid Robot Using a Convolutional Neural Network. IEEE Access 2022, 10, 89876–89886. [Google Scholar] [CrossRef]
  53. Chen, L.; Li, M.; Wu, M.; Pedrycz, W.; Hirota, K. Coupled Multimodal Emotional Feature Analysis Based on Broad-Deep Fusion Networks in Human–Robot Interaction. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 9663–9673. [Google Scholar] [CrossRef]
  54. Maehama, K.; Even, J.; Ishi, C.T.; Kanda, T. Enabling Robots to Distinguish Between Aggressive and Joking Attitudes. IEEE Robot. Autom. Lett. 2021, 6, 8037–8044. [Google Scholar] [CrossRef]
  55. Rawal, N.; Koert, D.; Turan, C.; Kersting, K.; Peters, J.; Stock-Homburg, R. Exgennet: Learning to Generate Robotic Facial Expression Using Facial Expression Recognition. Front. Robot. AI 2022, 8, 730317. [Google Scholar] [CrossRef] [PubMed]
  56. Khashman, A. Application of an Emotional Neural Network to Facial Recognition. Neural Comput. Appl. 2009, 18, 309–320. [Google Scholar] [CrossRef]
  57. Li, T.-H.S.; Kuo, P.-H.; Tsai, T.-N.; Luan, P.-C. CNN and LSTM Based Facial Expression Analysis Model for a Humanoid Robot. IEEE Access 2019, 7, 93998–94011. [Google Scholar] [CrossRef]
  58. Li, Y.K.; Meng, Q.H.; Wang, Y.X.; Hou, H.R. MMFN: Emotion Recognition by Fusing Touch Gesture and Facial Expression Information. Expert Syst. Appl. 2023, 228, 120469. [Google Scholar] [CrossRef]
  59. Daems, A.; Verfaillie, K. Viewpoint-Dependent Priming Effects in the Perception of Human Actions and Body Postures. Vis. Cogn. 1999, 6, 665–693. [Google Scholar] [CrossRef]
  60. Yagi, S.; Nakata, Y.; Nakamura, Y.; Ishiguro, H. Perception of Emotional Expression of Mobile Humanoid Robot Using Gait-Induced Upper Body Motion. IEEE Access 2021, 9, 124793–124804. [Google Scholar] [CrossRef]
  61. Mahzoon, H.; Ueda, A.; Yoshikawa, Y.; Ishiguro, H. Effect of Robot’s Vertical Body Movement on Its Perceived Emotion: A Preliminary Study on Vertical Oscillation and Transition. PLoS ONE 2022, 17, e0271789. [Google Scholar] [CrossRef] [PubMed]
  62. Papenmeier, F.; Uhrig, M.; Kirsch, A. Human Understanding of Robot Motion: The Role of Velocity and Orientation. Int. J. Soc. Robot. 2019, 11, 75–88. [Google Scholar] [CrossRef]
  63. Hirayama, T.; Okada, Y.; Kimoto, M.; Iio, T.; Shimohara, K.; Shiomi, M. Speed Effects in Touching Behaviours: Impact on Perceived Relationships in Robot-Robot Interactions. Adv. Robot. 2024, 38, 492–509. [Google Scholar] [CrossRef]
Figure 1. The block scheme for the main procedure of this study [26,32].
Figure 1. The block scheme for the main procedure of this study [26,32].
Sensors 24 07227 g001
Figure 2. The emotional body movements of human from the works of Zhang et al. and De Gelder et al., respectively, and the corresponding imitations by Yanshee robot. The behaviors shown from (af) are with emotions happiness, fear, anger, disgust, sadness, and surprise, respectively.
Figure 2. The emotional body movements of human from the works of Zhang et al. and De Gelder et al., respectively, and the corresponding imitations by Yanshee robot. The behaviors shown from (af) are with emotions happiness, fear, anger, disgust, sadness, and surprise, respectively.
Sensors 24 07227 g002
Figure 3. Scenario of a student watching video V41.
Figure 3. Scenario of a student watching video V41.
Sensors 24 07227 g003
Figure 4. (a) The scores of 72 robot body movements in six emotional categories. (b) The recognition rates of 72 robot body movements in six emotional categories.
Figure 4. (a) The scores of 72 robot body movements in six emotional categories. (b) The recognition rates of 72 robot body movements in six emotional categories.
Sensors 24 07227 g004
Figure 5. The neural network architecture of the proposed model.
Figure 5. The neural network architecture of the proposed model.
Sensors 24 07227 g005
Figure 6. The body parts of Yanshee robot.
Figure 6. The body parts of Yanshee robot.
Sensors 24 07227 g006
Figure 7. The body movements in V4 of Yanshee robot.
Figure 7. The body movements in V4 of Yanshee robot.
Sensors 24 07227 g007
Figure 8. The MSEs with the number of hidden neurons.
Figure 8. The MSEs with the number of hidden neurons.
Sensors 24 07227 g008
Figure 9. The MSEs of proposed BP model.
Figure 9. The MSEs of proposed BP model.
Sensors 24 07227 g009
Figure 10. The main flow of PSO-BP-RMSProp model.
Figure 10. The main flow of PSO-BP-RMSProp model.
Sensors 24 07227 g010
Figure 11. From (ac), the MSEs of PSO-BP model, BP-RMSProp model and PSO-BP-RMSProp model with the epochs, respectively.
Figure 11. From (ac), the MSEs of PSO-BP model, BP-RMSProp model and PSO-BP-RMSProp model with the epochs, respectively.
Sensors 24 07227 g011
Figure 12. From (ad), the data of emotion recognition obtained by proposed BP model, PSO-BP model, BP-RMSProp model and PSO-BP-RMSProp model, respectively.
Figure 12. From (ad), the data of emotion recognition obtained by proposed BP model, PSO-BP model, BP-RMSProp model and PSO-BP-RMSProp model, respectively.
Sensors 24 07227 g012
Figure 13. (a) The accuracy curve for each epoch of the CNN method. (b) The loss curve for each epoch of the CNN method.
Figure 13. (a) The accuracy curve for each epoch of the CNN method. (b) The loss curve for each epoch of the CNN method.
Sensors 24 07227 g013
Table 1. The Spearman correlation coefficients between the recognition rates and the scores of robot’s body movements.
Table 1. The Spearman correlation coefficients between the recognition rates and the scores of robot’s body movements.
Correlation CoefficientHappinessFearAngerDisgustSadnessSurprise
s0.970.790.8760.6740.8110.569
Table 2. The robot emotional body movements after screening.
Table 2. The robot emotional body movements after screening.
EmotionVideoEmotion Recognition Rate (%)M ValueSD
HappinessV486.275.161.63
V3986.275.841.60
V874.515.241.89
V2872.555.101.73
V2170.595.121.76
V2068.635.002.38
V364.714.121.70
V2262.754.431.93
V7160.784.061.95
V4460.784.061.99
V3158.824.081.79
V5558.824.392.30
V1558.824.612.01
V4054.904.512.12
V5750.984.021.97
V6750.983.982.20
V1650.984.102.08
AngerV1262.754.711.71
SadnessV6566.675.631.77
V6062.755.391.96
V5954.904.612.07
V6152.944.651.86
SurpriseV3358.824.551.74
V5052.944.761.91
V4952.944.631.73
Table 3. The motion state set for inputs and the emotion state set for outputs.
Table 3. The motion state set for inputs and the emotion state set for outputs.
The Motion State Set XnThe Emotion State Set Yn
Body Part Moving Direction Moving Mode
HeadLeftRotationHappiness
TorsoRightSwingFear
Right foot ForwardVerticalAnger
Left footBackwardExtendDisgust
Right handUpwardClose upSadness
Left handDownwardBendSurprise
Inward
Outward
Table 4. Some examples of the encoded inputs and outputs for modeling proposed neural network.
Table 4. Some examples of the encoded inputs and outputs for modeling proposed neural network.
VideoThe Encoded InputThe Encoded Output
V4{1, 1, 0, 1, 1, 1, 1, 0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0, 0}{1, 0, 0, 0, 0, 0}
V12{0, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0}{0, 0, 1, 0, 0, 0}
V33{0, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 1}{0, 0, 0, 0, 0, 1}
V65{0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 0, 0, 0, 0, 1, 1}{0, 0, 0, 0, 1, 0}
Table 5. Comparisons of emotion recognition results by different methods.
Table 5. Comparisons of emotion recognition results by different methods.
VideoMethod I *Method II *Method III *
EBMR-BPPSO-BPBP-RMSPropPSO-BP-RMSProp
V15HappinessHappinessHappinessHappinessHappinessHappiness
V16HappinessHappinessHappinessSadnessHappinessHappiness
V40HappinessHappinessHappinessHappinessHappinessHappiness
V49SurpriseHappinessHappinessHappinessSurpriseHappiness
V57HappinessHappinessHappinessHappinessHappinessHappiness
V67HappinessAngerSadnessAngerAngerHappiness
B8AngerAngerHappinessDisgustAngerHappiness
B9SadnessSadnessSadnessHappinessSadnessHappiness
B14FearHappinessHappinessFearFearHappiness
* Methods I, II and III represent the subjective emotion evaluation, the emotion recognition by proposed models and the emotion recognition by CNN, respectively.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, W.; Jiang, T.; Zhai, W.; Zha, F. An Emotion Recognition Method for Humanoid Robot Body Movements Based on a PSO-BP-RMSProp Neural Network. Sensors 2024, 24, 7227. https://doi.org/10.3390/s24227227

AMA Style

Gao W, Jiang T, Zhai W, Zha F. An Emotion Recognition Method for Humanoid Robot Body Movements Based on a PSO-BP-RMSProp Neural Network. Sensors. 2024; 24(22):7227. https://doi.org/10.3390/s24227227

Chicago/Turabian Style

Gao, Wa, Tanfeng Jiang, Wanli Zhai, and Fusheng Zha. 2024. "An Emotion Recognition Method for Humanoid Robot Body Movements Based on a PSO-BP-RMSProp Neural Network" Sensors 24, no. 22: 7227. https://doi.org/10.3390/s24227227

APA Style

Gao, W., Jiang, T., Zhai, W., & Zha, F. (2024). An Emotion Recognition Method for Humanoid Robot Body Movements Based on a PSO-BP-RMSProp Neural Network. Sensors, 24(22), 7227. https://doi.org/10.3390/s24227227

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop