TY - GEN
T1 - Multimodal classification of interruptions in humans' interaction
AU - Yang, Liu
AU - Achard, Catherine
AU - Pelachaud, Catherine
N1 - Publisher Copyright:
© 2022 ACM.
PY - 2022/11/7
Y1 - 2022/11/7
N2 - During an interaction interruptions occur frequently. Interruptions may arise to fulfill different goals such as changing the topic of conversation abruptly, asking for clarification, completing the current speaker's turn. Interruptions may be cooperative or competitive depending on the interrupter's intention. Our main goal is to endow a Socially Interactive Agent with the capacity to handle user interruptions in dyadic interaction. It requires the agent to detect an interruption and recognize its type (cooperative/competitive), and then to plan its behaviours to respond appropriately. As a first step towards this goal, we developed a multimodal classification model using acoustic features, facial expression, head movement, and gaze direction from both, the interrupter and the interruptee. The classification model learns from the sequential information to automatically identify interruptions type. We also present studies we conducted to measure the shortest delay needed (0.6s) for our classification model to identify interruption types with a high classification accuracy (81%). On average, most interruption overlaps last longer than 0.6s, so a Socially Interactive Agent has time to detect and recognize an interruption type and can respond in a timely manner to its human interlocutor's interruption.
AB - During an interaction interruptions occur frequently. Interruptions may arise to fulfill different goals such as changing the topic of conversation abruptly, asking for clarification, completing the current speaker's turn. Interruptions may be cooperative or competitive depending on the interrupter's intention. Our main goal is to endow a Socially Interactive Agent with the capacity to handle user interruptions in dyadic interaction. It requires the agent to detect an interruption and recognize its type (cooperative/competitive), and then to plan its behaviours to respond appropriately. As a first step towards this goal, we developed a multimodal classification model using acoustic features, facial expression, head movement, and gaze direction from both, the interrupter and the interruptee. The classification model learns from the sequential information to automatically identify interruptions type. We also present studies we conducted to measure the shortest delay needed (0.6s) for our classification model to identify interruption types with a high classification accuracy (81%). On average, most interruption overlaps last longer than 0.6s, so a Socially Interactive Agent has time to detect and recognize an interruption type and can respond in a timely manner to its human interlocutor's interruption.
KW - Interruption Classification
KW - multimodality
KW - Nonverbal Behaviour
KW - Socially Interactive Agent
UR - http://www.scopus.com/inward/record.url?scp=85142824371&partnerID=8YFLogxK
U2 - 10.1145/3536221.3556604
DO - 10.1145/3536221.3556604
M3 - Conference Proceeding
AN - SCOPUS:85142824371
T3 - ACM International Conference Proceeding Series
SP - 597
EP - 604
BT - ICMI 2022 - Proceedings of the 2022 International Conference on Multimodal Interaction
PB - Association for Computing Machinery
T2 - 24th ACM International Conference on Multimodal Interaction, ICMI 2022
Y2 - 7 November 2022 through 11 November 2022
ER -