Projects per year
Abstract
To be able to react to interaction ruptures such as errors, a robot needs a way of realizing such a rupture occurred. We test whether it is possible to detect interaction ruptures from the user's anonymized speech, posture, and facial features. We showcase how to approach this task, presenting a time series classification pipeline that works well with various machine learning models. A sliding window is applied to the data and the continuously updated predictions make it suitable for detecting ruptures in real-time. Our best model, an ensemble of MiniRocket classifiers, is the winning approach to the ICMI ERR@HRI challenge. A feature importance analysis shows that the model heavily relies on speaker diarization data that indicates who spoke when. Posture data, on the other hand, impedes performance. Our code is available online (https://github.com/lwachowiak/HRI-Error-Detection-STAI).
Original language | English |
---|---|
Title of host publication | Proceedings of the 26th International Conference on Multimodal Interaction (ICMI ’24) |
Publisher | ACM |
Number of pages | 9 |
DOIs | |
Publication status | Accepted/In press - 10 Aug 2024 |
Keywords
- hri
- robotics
- machine learning
- time series
- multimodal
Fingerprint
Dive into the research topics of 'A Time Series Classification Pipeline for Detecting Interaction Ruptures in HRI Based on User Reactions'. Together they form a unique fingerprint.Projects
- 2 Active
-
-
COHERENT - COllaborative HiErarchical Robotic ExplaNaTions
EPSRC Engineering and Physical Sciences Research Council
1/04/2021 → 31/03/2025
Project: Research