Publication: Audio-visual prediction of head-nod and turn-taking events in dyadic interactions
Program
KU Authors
Co-Authors
Advisor
Publication Date
2018
Language
English
Type
Conference proceeding
Journal Title
Journal ISSN
Volume Title
Abstract
Head-nods and turn-taking both significantly contribute conversational dynamics in dyadic interactions. Timely prediction and use of these events is quite valuable for dialog management systems in human-robot interaction. in this study, we present an audio-visual prediction framework for the head-nod and turn taking events that can also be utilized in real-time systems. Prediction systems based on Support vector Machines (SVM) and Long Short-Term Memory Recurrent Neural Networks (LSTM-RNN) are trained on human-human conversational data. Unimodal and multi-modal classification performances of head-nod and turn-taking events are reported over the IEMOCaP dataset.
Description
Source:
19th Annual Conference of the international Speech Communication Association (interspeech 2018), Vols 1-6: Speech Research for Emerging Markets in Multilingual Societies
Publisher:
Isca-int Speech Communication assoc
Keywords:
Subject
Computer Science, Artificial intelligence, Electrical electronics engineering