Next Article in Journal
Human Attitudes in Robotic Path Programming: A Pilot Study of User Experience in Manual and XR-Controlled Robotic Arm Manipulation
Next Article in Special Issue
A Systematic Review on Artificial Intelligence-Based Multimodal Dialogue Systems Capable of Emotion Recognition
Previous Article in Journal
Self-Created Film as a Resource in a Multimodal Conversational Narrative
Previous Article in Special Issue
Choice Vectors: Streamlining Personal AI Alignment Through Binary Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

diaLogic: A Multi-Modal Framework for Automated Team Behavior Modeling Based on Speech Acquisition

Department of Electrical and Computer Engineering, Stony Brook University, Stony Brook, NY 11794-2350, USA
*
Author to whom correspondence should be addressed.
Multimodal Technol. Interact. 2025, 9(3), 26; https://doi.org/10.3390/mti9030026
Submission received: 25 December 2024 / Revised: 12 February 2025 / Accepted: 6 March 2025 / Published: 10 March 2025

Abstract

This paper presents diaLogic, a humans-in-the-loop system for modeling the behavior of teams during collective problem solving. Team behavior is modeled using multi-modal data about cognition, social interactions, and emotions acquired from speech inputs. The system includes methods for speaker diarization, speaker interaction characterization, speaker emotion recognition, and speech-to-text conversion. Hypotheses about the invariant and differentiated aspects of teams are extracted using the similarities and dissimilarities of their behavior over time. Hypothesis extraction, a novel contribution of this work, uses a method to identify the clauses and concepts in each spoken sentence. Experiments present system performance for a broad set of cases of team behavior during problem solving. The average errors of the various methods are between 6% and 21%. The system can be used in a broad range of applications, from education to team research and therapy.
Keywords: multi-modal data collection; speaker diarization; speaker interactions; speaker emotion recognition; hypothesis extraction; humans in the loop multi-modal data collection; speaker diarization; speaker interactions; speaker emotion recognition; hypothesis extraction; humans in the loop

Share and Cite

MDPI and ACS Style

Duke, R.; Doboli, A. diaLogic: A Multi-Modal Framework for Automated Team Behavior Modeling Based on Speech Acquisition. Multimodal Technol. Interact. 2025, 9, 26. https://doi.org/10.3390/mti9030026

AMA Style

Duke R, Doboli A. diaLogic: A Multi-Modal Framework for Automated Team Behavior Modeling Based on Speech Acquisition. Multimodal Technologies and Interaction. 2025; 9(3):26. https://doi.org/10.3390/mti9030026

Chicago/Turabian Style

Duke, Ryan, and Alex Doboli. 2025. "diaLogic: A Multi-Modal Framework for Automated Team Behavior Modeling Based on Speech Acquisition" Multimodal Technologies and Interaction 9, no. 3: 26. https://doi.org/10.3390/mti9030026

APA Style

Duke, R., & Doboli, A. (2025). diaLogic: A Multi-Modal Framework for Automated Team Behavior Modeling Based on Speech Acquisition. Multimodal Technologies and Interaction, 9(3), 26. https://doi.org/10.3390/mti9030026

Article Metrics

Back to TopTop