Publication:
Audio-visual prediction of head-nod and turn-taking events in dyadic interactions

Placeholder

Organizational Units

Program

KU Authors

Co-Authors

Advisor

Publication Date

2018

Language

English

Type

Conference proceeding

Journal Title

Journal ISSN

Volume Title

Abstract

Head-nods and turn-taking both significantly contribute conversational dynamics in dyadic interactions. Timely prediction and use of these events is quite valuable for dialog management systems in human-robot interaction. in this study, we present an audio-visual prediction framework for the head-nod and turn taking events that can also be utilized in real-time systems. Prediction systems based on Support vector Machines (SVM) and Long Short-Term Memory Recurrent Neural Networks (LSTM-RNN) are trained on human-human conversational data. Unimodal and multi-modal classification performances of head-nod and turn-taking events are reported over the IEMOCaP dataset.

Description

Source:

19th Annual Conference of the international Speech Communication Association (interspeech 2018), Vols 1-6: Speech Research for Emerging Markets in Multilingual Societies

Publisher:

Isca-int Speech Communication assoc

Keywords:

Subject

Computer Science, Artificial intelligence, Electrical electronics engineering

Citation

Endorsement

Review

Supplemented By

Referenced By

Copy Rights Note

0

Views

0

Downloads

View PlumX Details