The Development of an Emotional Embodied Conversational Agent and the Evaluation of the Effect of Response Delay on User Impression
Abstract
:1. Introduction
2. Related Work
- Recognizing and responding to verbal and nonverbal input;
- Generating verbal and nonverbal output;
- Dealing with conversational functions, such as turn-taking, feedback, and repair mechanisms;
- Giving signals that indicate the state of the conversation, as well as contributing new propositions to the discourse.
3. Development of an Embodied Emotional Conversational Agent
3.1. Overview
3.2. Emotion Recognition from Facial Image
3.3. Speech Recognition and Emotion Estimation
3.4. Dialogue Management
3.5. Generation of Face Image and Facial Expressions
- Prioritizing the answer’s emotion: The primary factor is the emotion predicted from the generated answer. If the prediction emotion is not neutral, the ECA’s face is rendered using that emotion.
- Maintaining Current Expression: In instances where the predicted emotion of the answer is neutral, the agent keeps the same facial expression.
- Reflecting the user’s emotion: If the agent’s facial expression is neutral twice, the agent mimics the user’s facial expression. It mimics the behavior of entrainment, where individuals can adjust their speaking rate, pitch, and gesture to match those of their conversation partner [47]. In the context of an ECA, entrainment can foster more engagement for the human user [48].
3.6. Text-to-Speech and Lip Syncing
3.7. Facial Animation Rendering
3.8. Controlling the Entire System
4. Experiment
4.1. Overview
4.2. Experimental Conditions
- We explained the purpose and procedure of the experiment to the participant.
- We asked the participant for a pre-dialogue questionnaire (Table 6).
- The participant held four rounds of dialogue with the agent. In each round, a participant conversed with the ECA about one particular artwork.
- After each conversation, the participant answered the questionnaire (Table 7) evaluating the current dialogue.
- After four rounds of dialogue, the participant answered the post-dialogue questionnaire (Table 8).
4.3. Questionnaire
4.4. Number of Utterances in a Dialogue
4.5. The Real Response Delay
4.6. Total Impression of the Conversation by the System
4.7. Effect of Response Delay and Emotional Expression
4.8. Effect of Demographic Characteristics
4.9. Questionnaire Results After Four Rounds of Conversation
5. Conclusions
Supplementary Materials
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Cassell, J. Embodied conversational interface agents. Commun. ACM 2000, 43, 70–78. [Google Scholar] [CrossRef]
- Cassell, J. Embodied conversational agents: Representation and intelligence in user interfaces. AI Mag. 2001, 22, 67. [Google Scholar]
- Allouch, M.; Azaria, A.; Azoulay, R. Conversational agents: Goals, technologies, vision and challenges. Sensors 2021, 21, 8448. [Google Scholar] [CrossRef]
- Santos-Perez, M.; Gonzalez-Parada, E.; Cano-Garcia, J.M. Mobile embodied conversational agent for task specific applications. IEEE Trans. Consum. Electron. 2013, 59, 610–614. [Google Scholar] [CrossRef]
- Miyake, S.; Ito, A. A spoken dialogue system using virtual conversational agent with augmented reality. In Proceedings of the 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA-ASC), Hollywood, CA, USA, 3–6 December 2012; IEEE: New York, NY, USA, 2012; pp. 1–4. [Google Scholar]
- Fontecha, J.; González, I.; Salas-Seguín, A. Using Conversational Assistants and Connected Devices to Promote a Responsible Energy Consumption at Home. Proceedings 2019, 31, 32. [Google Scholar] [CrossRef]
- André, E. Design and evaluation of embodied conversational agents for educational and advisory software. In Handbook of Conversation Design for Instructional Applications; IGI Global: Hershey, PA, USA, 2008; pp. 343–362. [Google Scholar]
- Craig, P.; Roa-Seïler, N.; Rosano, F.; Díaz, M. The role of embodied conversational agents in collaborative face to face computer supported learning games. In Proceedings of the 26th International Conference on System Research, Informatics & Cybernetics, Baden-Baden, Germany, 29 July–2 August 2013. [Google Scholar]
- Sebastian, J.; Richards, D. Changing stigmatizing attitudes to mental health via education and contact with embodied conversational agents. Comput. Hum. Behav. 2017, 73, 479–488. [Google Scholar] [CrossRef]
- Laranjo, L.; Dunn, A.G.; Tong, H.L.; Kocaballi, A.B.; Chen, J.; Bashir, R.; Surian, D.; Gallego, B.; Magrabi, F.; Lau, A.Y.; et al. Conversational agents in healthcare: A systematic review. J. Am. Med. Inform. Assoc. 2018, 25, 1248–1258. [Google Scholar] [CrossRef]
- Bin Sawad, A.; Narayan, B.; Alnefaie, A.; Maqbool, A.; Mckie, I.; Smith, J.; Yuksel, B.; Puthal, D.; Prasad, M.; Kocaballi, A.B. A systematic review on healthcare artificial intelligent conversational agents for chronic conditions. Sensors 2022, 22, 2625. [Google Scholar] [CrossRef]
- Doumanis, I.; Smith, S. Evaluating the impact of embodied conversational agents (ECAs) attentional behaviors on user retention of cultural content in a simulated mobile environment. In Proceedings of the 7th Workshop on Eye Gaze in Intelligent Human Machine Interaction: Eye-Gaze & Multimodality, Istanbul, Turkey, 16 November 2014; pp. 27–32. [Google Scholar]
- Pelachaud, C.; Bilvi, M. Computational Model of Believable Conversational Agents. In Communication in Multiagent Systems: Agent Communication Languages and Conversation Policies; Huget, M.P., Ed.; Springer: Berlin/Heidelberg, Germany, 2003; pp. 300–317. [Google Scholar] [CrossRef]
- Wirzberger, M.; Schmidt, R.; Georgi, M.; Hardt, W.; Brunnett, G.; Rey, G.D. Effects of system response delays on elderly humans’ cognitive performance in a virtual training scenario. Sci. Rep. 2019, 9, 8291. [Google Scholar] [CrossRef]
- Strömbergsson, S.; Hjalmarsson, A.; Edlund, J.; House, D. Timing responses to questions in dialogue. In Proceedings of the Interspeech, Lyon, France, 25–29 August 2013; pp. 2584–2588. [Google Scholar] [CrossRef]
- Sacks, H.; Schegloff, E.A.; Jefferson, G. A simplest systematics for the organization of turn-taking for conversation. Language 1974, 50, 696–735. [Google Scholar] [CrossRef]
- Hara, K.; Inoue, K.; Takanashi, K.; Kawahara, T. Turn-Taking Prediction Based on Detection of Transition Relevance Place. In Proceedings of the Interspeech, Graz, Austria, 15–19 September 2019; pp. 4170–4174. [Google Scholar]
- Yahagi, R.; Chiba, Y.; Nose, T.; Ito, A. Multimodal Dialogue Response Timing Estimation Using Dialogue Context Encoder. In Conversational AI for Natural Human-Centric Interaction, Proceedings of the 12th International Workshop on Spoken Dialogue System Technology, IWSDS 2021, Singapore, 15–17 November 2021; Springer: Berlin/Heidelberg, Germany, 2022; pp. 133–141. [Google Scholar]
- Sakuma, J.; Fujie, S.; Zhao, H.; Kobayashi, T. Improving the response timing estimation for spoken dialogue systems by reducing the effect of speech recognition delay. In Proceedings of the Interspeech, Dublin, Ireland, 20–24 August 2023; Volume 2023, pp. 2668–2672. [Google Scholar]
- Trimboli, C.; Walker, M.B. Switching pauses in cooperative and competitive conversations. J. Exp. Soc. Psychol. 1984, 20, 297–311. [Google Scholar] [CrossRef]
- Devillers, L.; Vasilescu, I.; Vidrascu, L. F0 and pause features analysis for anger and fear detection in real-life spoken dialogs. In Proceedings of the Speech Prosody, Nara, Japan, 23–26 March 2004. [Google Scholar]
- Cassell, J.; Bickmore, T.; Billinghurst, M.; Campbell, L.; Chang, K.; Vilhjálmsson, H.; Yan, H. Embodiment in conversational interfaces: Rea. In Proceedings of the the SIGCHI Conference on Human Factors in Computing Systems, CHI ’99, New York, NY, USA, 15–20 May 1999; pp. 520–527. [Google Scholar] [CrossRef]
- Van Pinxteren, M.M.; Pluymaekers, M.; Lemmink, J.G. Human-like communication in conversational agents: A literature review and research agenda. J. Serv. Manag. 2020, 31, 203–225. [Google Scholar] [CrossRef]
- Diederich, S.; Brendel, A.B.; Morana, S.; Kolbe, L. On the design of and interaction with conversational agents: An organizing and assessing review of human-computer interaction research. J. Assoc. Inf. Syst. 2022, 23, 96–138. [Google Scholar] [CrossRef]
- Cassell, J.; Thorisson, K.R. The power of a nod and a glance: Envelope vs. emotional feedback in animated conversational agents. Appl. Artif. Intell. 1999, 13, 519–538. [Google Scholar] [CrossRef]
- Ochs, M.; Pelachaud, C.; Sadek, D. An empathic virtual dialog agent to improve human-machine interaction. In Proceedings of the 7th International Joint Conference on Autonomous Agents and Multiagent Systems, Estoril, Portugal, 12–16 May 2008; Volume 1, pp. 89–96. [Google Scholar]
- Becker, C.; Kopp, S.; Wachsmuth, I. Simulating the emotion dynamics of a multimodal conversational agent. In Proceedings of the Tutorial and Research Workshop on Affective Dialogue Systems, Kloster Irsee, Germany, 14–16 June 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 154–165. [Google Scholar]
- Egges, A.; Kshirsagar, S.; Magnenat-Thalmann, N. Generic personality and emotion simulation for conversational agents. Comput. Animat. Virtual Worlds 2004, 15, 1–13. [Google Scholar] [CrossRef]
- Ochs, M.; Sadek, D.; Pelachaud, C. A formal model of emotions for an empathic rational dialog agent. Auton. Agents -Multi-Agent Syst. 2012, 24, 410–440. [Google Scholar] [CrossRef]
- Chiba, Y.; Nose, T.; Kase, T.; Yamanaka, M.; Ito, A. An analysis of the effect of emotional speech synthesis on non-task-oriented dialogue system. In Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue, Melbourne, Australia, 12–14 July 2018; pp. 371–375. [Google Scholar]
- Yamanaka, M.; Chiba, Y.; Nose, T.; Ito, A. A study on a spoken dialogue system with cooperative emotional speech synthesis using acoustic and linguistic information. In Recent Advances in Intelligent Information Hiding and Multimedia Signal Processing: Proceeding of the Fourteenth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, Sendai, Japan, 26–28 November 2018; Springer: Berlin/Heidelberg, Germany, 2019; Volume 2, pp. 101–108. [Google Scholar]
- Firdaus, M.; Chauhan, H.; Ekbal, A.; Bhattacharyya, P. EmoSen: Generating sentiment and emotion controlled responses in a multimodal dialogue system. IEEE Trans. Affect. Comput. 2020, 13, 1555–1566. [Google Scholar] [CrossRef]
- Loveys, K.; Sagar, M.; Broadbent, E. The effect of multimodal emotional expression on responses to a digital human during a self-disclosure conversation: A computational analysis of user language. J. Med. Syst. 2020, 44, 143. [Google Scholar] [CrossRef]
- Saha, T.; Reddy, S.; Das, A.; Saha, S.; Bhattacharyya, P. A shoulder to cry on: Towards a motivational virtual assistant for assuaging mental agony. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, WA, USA, 10–15 July 2022; pp. 2436–2449. [Google Scholar]
- Jokinen, K.; Homma, K.; Matsumoto, Y.; Fukuda, K. Integration and interaction of trustworthy ai in a virtual coach—An overview of EU-Japan collaboration on eldercare. In Proceedings of the Annual Conference of the Japanese Society for Artificial Intelligence, Yokohama, Japan, 13–15 November 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 190–200. [Google Scholar]
- Pauw, L.S.; Sauter, D.A.; van Kleef, G.A.; Lucas, G.M.; Gratch, J.; Fischer, A.H. The avatar will see you now: Support from a virtual human provides socio-emotional benefits. Comput. Hum. Behav. 2022, 136, 107368. [Google Scholar] [CrossRef]
- Funk, M.; Cunningham, C.; Kanver, D.; Saikalis, C.; Pansare, R. Usable and acceptable response delays of conversational agents in automotive user interfaces. In Proceedings of the 12th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Virtual, 21–22 September 2020; pp. 262–269. [Google Scholar]
- Gnewuch, U.; Morana, S.; Adam, M.; Maedche, A. Faster is not always better: Understanding the effect of dynamic response delays in human-chatbot interaction. In Proceedings of the European Conference on Information Systems, Portsmouth, UK, 23–28 June 2018; p. 113. [Google Scholar]
- Heeman, P.; Lunsford, R. Turn-Taking Offsets and Dialogue Context. In Proceedings of the Interspeech, Stockholm, Sweden, 20–24 August 2017; Volume 2017, pp. 1671–1675. [Google Scholar] [CrossRef]
- Jolibois, S.; Ito, A.; Nose, T. Multimodal Expressive Embodied Conversational Agent Design. In Proceedings of the International Conference on Human-Computer Interaction, Copenhagen, Denmark, 23–28 July 2023; Springer: Cham, Switzerland, 2023; pp. 244–249. [Google Scholar]
- Baltrušaitis, T.; Robinson, P.; Morency, L.P. Openface: An open source facial behavior analysis toolkit. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; IEEE: New York, NY, USA, 2016; pp. 1–10. [Google Scholar]
- Ekman, P.; Friesen, W.V. Facial Action Coding System: A Technique for the Measurement of Facial Movement; Consulting Psychologists Press: Palo Alto, CA, USA, 1978. [Google Scholar]
- Lucey, P.; Cohn, J.F.; Kanade, T.; Saragih, J.; Ambadar, Z.; Matthews, I. The extended Cohn-Kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, San Francisco, CA, USA, 13–18 June 2010; IEEE: New York, NY, USA, 2010; pp. 94–101. [Google Scholar]
- Demszky, D.; Movshovitz-Attias, D.; Ko, J.; Cowen, A.; Nemade, G.; Ravi, S. GoEmotions: A Dataset of Fine-Grained Emotions. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Virtual, 5–10 July 2020; pp. 4040–4054. [Google Scholar]
- Nisimura, R.; Lee, A.; Saruwatari, H.; Shikano, K. Public speech-oriented guidance system with adult and child discrimination capability. In Proceedings of the 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing, Montreal, QC, Canada, 17–21 May 2004; IEEE: New York, NY, USA, 2004; Volume 1, pp. 433–436. [Google Scholar]
- Inoue, R.; Kurosawa, Y.; Mera, K.; Takezawa, T. A question-and-answer classification technique for constructing and managing spoken dialog system. In Proceedings of the 2011 International Conference on Speech Database and Assessments (Oriental COCOSDA), Hsinchu, Taiwan, 26–28 October 2011; IEEE: New York, NY, USA, 2011; pp. 97–101. [Google Scholar]
- Navarretta, C. Mirroring facial expressions and emotions in dyadic conversations. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), Portorož, Slovenia, 23–28 May 2016; pp. 469–474. [Google Scholar]
- Stevens, C.J.; Pinchbeck, B.; Lewis, T.; Luerssen, M.; Pfitzner, D.; Powers, D.M.; Abrahamyan, A.; Leung, Y.; Gibert, G. Mimicry and expressiveness of an ECA in human-agent interaction: Familiarity breeds content! Comput. Cogn. Sci. 2016, 2, 1–14. [Google Scholar] [CrossRef]
- Miura, I. Switching pauses in adult-adult and child-child turn takings: An initial study. J. Psycholinguist. Res. 1993, 22, 383–395. [Google Scholar] [CrossRef]
- Mori, H. An analysis of switching pause duration as a paralinguistic feature in expressive dialogues. Acoust. Sci. Technol. 2009, 30, 376–378. [Google Scholar] [CrossRef]
- Peras, D. Chatbot evaluation metrics. In Economic and Social Development: Book of Proceedings; Varazdin Development and Entrepreneurship Agency: Varazdin, Croatia, 2018; pp. 89–97. [Google Scholar]
- Oosterhof, N.N.; Todorov, A. Shared perceptual basis of emotional expressions and trustworthiness impressions from faces. Emotion 2009, 9, 128. [Google Scholar] [CrossRef] [PubMed]
- Chen, X.; Zhou, M.; Wang, R.; Pan, Y.; Mi, J.; Tong, H.; Guan, D. Evaluating Response Delay of Multimodal Interface in Smart Device. In Design, User Experience, and Usability. Practice and Case Studies; Marcus, A., Wang, W., Eds.; Springer: Cham, Switzerland, 2019; pp. 408–419. [Google Scholar]
- Asaka, S.; Itoyama, K.; Nakadai, K. Improving Impressions of Response Delay in AI-based Spoken Dialogue Systems. In Proceedings of the 2024 33rd IEEE International Conference on Robot and Human Interactive Communication (ROMAN), Pasadena, CA, USA, 26–30 August 2024; IEEE: New York, NY, USA, 2024; pp. 1416–1421. [Google Scholar]
No. | Description | No. | Description |
---|---|---|---|
01 | Inner Brow Raiser | 14 | Dimpler |
02 | Outer Brow Raiser | 15 | Lip Corner Depressor |
04 | Brow Lowerer | 17 | Chin Raiser |
05 | Upper Lid Raiser | 20 | Lip Stretcher |
06 | Cheek Raiser | 23 | Lip Tightener |
07 | Lid Tightener | 25 | Lips Part |
09 | Nose Wrinkler | 26 | Jaw Drop |
10 | Upper Lip Raiser | 45 | Blink |
12 | Lip Corner Puller |
Number of layers/transformer blocks | 12 |
Number of units in the hidden layers | 768 |
Total number of parameters | 110 M |
Number of self-attention heads | 12 |
Number of epochs | 4 |
Batch size | 16 |
Learning rate | |
Validation split | 85/15 |
Emotion | Accuracy (%) |
---|---|
Anger | 46 |
Annoyance | 41 |
Confusion | 35 |
Curiosity | 45 |
Disgust | 37 |
Embarrassment | 37 |
Excitement | 21 |
Fear | 65 |
Grief | 6 |
Joy | 71 |
Nervousness | 25 |
Pride | 0 |
Sadness | 59 |
Surprise | 61 |
Weighted macro-average | 45 |
No. | Name of the Artwork | Painter |
---|---|---|
1 | A Hughenot on St. Bartholomew’s Day | John Everett Millais |
2 | Liberty Leading the People | Eugéne Delacroix |
3 | The Roses of Heliogabalus | Lawrence Alma-Tadema |
4 | The Starry Night | Vincent van Gogh |
No. | Item |
---|---|
1 | Title in English and original language |
2 | Place and date of creation |
3 | Dimensions |
4 | Genre |
5 | Format |
6 | Subject matter |
7 | Cultural and historical context in which the artwork was born |
8 | The intent of the author for creating the piece |
9 | The current conservation place and methods |
10 | Previous exhibitions the artwork was part of |
11 | The ownership and sale history |
12 | Technical analysis done on the artwork |
13 | The painter’s name |
14 | The painter’s biography |
15 | The painter’s artistic movement(s) |
16 | The painter’s famous artworks |
17 | Inspirations/references used for the artwork |
Q1 | Gender |
Q2 | Are you familiar with the concept of a chatbot? (No, I don’t know what a chatbot is./I don’t have much knowledge about it./Yes, I am familiar with the concept.) |
Q3 | Have you already used a chatbot? (Never used before, Seldom use, Frequently use) |
Q4 | How would you rate your English level? (Beginner, Intermediate, Advanced, Fluent) |
Q1 | Animation |
Q2 | Conversation skill of the chatbot (Flow of the conversation, phrasing) |
Q3 | Information accuracy (Does the agent’s answer correspond to your question?) |
Q4 | Interactivity (Is the agent responsive to your inputs?) |
Q5 | Interface (Layout, Visibility) |
Q6 | Naturalness of the conversation (Does the agent look human-like when conversing?) |
Q7 | Responsiveness (How fast does the agent answer back?) |
Q8 | Overall satisfaction |
Q9 | Usability (Ease of use) |
Q10 | Visual aspect (Graphics) |
Q1 | Among the following parameters, what do you think hindered your conversation with the chatbot the most? (Visual aspect, Animation, Conversation skill of the agent, Conversation topic, Interactivity, Responsiveness, Interface) |
Q2 | Please elaborate on your choice |
Q3 | Which emotions/expressions were overrepresented? Underrepresented? |
Q4 | Were you knowledgeable on the topic of artworks? Did the conversation topic hinder your ability to interact with the agent? |
Q5 | Do you think one of the aspects of the system is not exploited enough? |
Q6 | If you have any comments, please write them here |
Round 1 | Round 2 | Round 3 | Round 4 | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
No. | Art | RD | Emo | Art | RD | Emo | Art | RD | Emo | Art | RD | Emo |
1 | 1 | 500 | Yes | 2 | 1000 | No | 3 | 1500 | Yes | 4 | 2000 | No |
2 | 2 | 500 | No | 3 | 1000 | Yes | 4 | 2000 | No | 1 | 1500 | Yes |
3 | 3 | 500 | Yes | 4 | 1500 | No | 1 | 1000 | Yes | 2 | 2000 | No |
4 | 2 | 500 | No | 1 | 1500 | Yes | 4 | 2000 | No | 3 | 1000 | Yes |
5 | 1 | 500 | Yes | 2 | 2000 | No | 3 | 1000 | Yes | 4 | 1500 | No |
6 | 2 | 500 | No | 3 | 2000 | Yes | 4 | 1500 | No | 1 | 1000 | Yes |
7 | 3 | 1000 | Yes | 4 | 500 | No | 1 | 1500 | Yes | 2 | 2000 | No |
8 | 4 | 1000 | No | 1 | 500 | Yes | 2 | 2000 | No | 3 | 1500 | Yes |
9 | 1 | 1000 | Yes | 2 | 1500 | No | 3 | 2000 | Yes | 4 | 500 | No |
10 | 2 | 1000 | No | 3 | 1500 | Yes | 4 | 500 | No | 1 | 2000 | Yes |
11 | 3 | 1000 | Yes | 4 | 2000 | No | 1 | 500 | Yes | 2 | 1500 | No |
12 | 4 | 1000 | No | 1 | 2000 | Yes | 2 | 1500 | No | 3 | 500 | Yes |
13 | 1 | 1500 | Yes | 2 | 500 | No | 3 | 1000 | Yes | 4 | 2000 | No |
14 | 2 | 1500 | No | 3 | 500 | Yes | 4 | 2000 | No | 1 | 1000 | Yes |
15 | 3 | 1500 | Yes | 4 | 1000 | No | 1 | 500 | Yes | 2 | 2000 | No |
16 | 4 | 1500 | No | 1 | 1000 | Yes | 2 | 2000 | No | 3 | 500 | Yes |
17 | 1 | 1500 | Yes | 2 | 2000 | No | 3 | 1000 | Yes | 4 | 500 | No |
18 | 2 | 1500 | No | 3 | 2000 | Yes | 4 | 500 | No | 1 | 1000 | Yes |
19 | 3 | 2000 | Yes | 4 | 500 | No | 1 | 1000 | Yes | 2 | 1500 | No |
20 | 4 | 2000 | No | 1 | 500 | Yes | 2 | 1500 | No | 3 | 1000 | Yes |
21 | 1 | 2000 | Yes | 2 | 1000 | No | 3 | 1500 | Yes | 4 | 500 | No |
22 | 2 | 2000 | No | 3 | 1000 | Yes | 4 | 500 | No | 1 | 1500 | Yes |
23 | 3 | 1500 | Yes | 4 | 1000 | No | 1 | 500 | Yes | 2 | 2000 | No |
24 | 4 | 2000 | No | 1 | 1500 | Yes | 2 | 1000 | No | 3 | 500 | Yes |
25 | 2 | 2000 | No | 3 | 1000 | Yes | 4 | 500 | No | 1 | 1500 | Yes |
Variable | Df | Pillai’s Trace | Approx. F | Df1 | Df2 | p |
---|---|---|---|---|---|---|
Response delay | 3 | 0.23772 | 0.7315 | 30 | 255 | 0.84712 |
Emotion | 1 | 0.16986 | 1.6983 | 10 | 83 | 0.09470 |
Delay × Emotion | 3 | 0.45136 | 1.5053 | 30 | 255 | 0.04964 |
Item | Response Delay | Emotional Expression | Interaction |
---|---|---|---|
Animation | 0.917 | 0.298 | 0.093 |
Conversation skill | 0.794 | 0.192 | 0.270 |
Information accuracy | 0.529 | 0.027 * | 0.581 |
Interactivity | 0.119 | 0.237 | 0.287 |
Interface | 0.771 | 0.649 | 0.062 |
Naturalness | 0.336 | 0.474 | 0.683 |
Responsiveness | 0.026 * | 0.069 | 0.038 * |
Satisfaction | 0.938 | 0.027 * | 0.485 |
Usability | 0.781 | 0.383 | 0.122 |
Visual aspect | 0.834 | 0.273 | 0.258 |
English | Use of Chatbot | |
---|---|---|
No | Yes | |
Advanced | 7 | 8 |
Fluent | 1 | 4 |
Intermediate | 4 | 1 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jolibois, S.C.; Ito, A.; Nose, T. The Development of an Emotional Embodied Conversational Agent and the Evaluation of the Effect of Response Delay on User Impression. Appl. Sci. 2025, 15, 4256. https://doi.org/10.3390/app15084256
Jolibois SC, Ito A, Nose T. The Development of an Emotional Embodied Conversational Agent and the Evaluation of the Effect of Response Delay on User Impression. Applied Sciences. 2025; 15(8):4256. https://doi.org/10.3390/app15084256
Chicago/Turabian StyleJolibois, Simon Christophe, Akinori Ito, and Takashi Nose. 2025. "The Development of an Emotional Embodied Conversational Agent and the Evaluation of the Effect of Response Delay on User Impression" Applied Sciences 15, no. 8: 4256. https://doi.org/10.3390/app15084256
APA StyleJolibois, S. C., Ito, A., & Nose, T. (2025). The Development of an Emotional Embodied Conversational Agent and the Evaluation of the Effect of Response Delay on User Impression. Applied Sciences, 15(8), 4256. https://doi.org/10.3390/app15084256