作者
Ioannis Kansizoglou, Loukas Bampis, Antonios Gasteratos
发表日期
2019/12/20
期刊
IEEE Transactions on Affective Computing
卷号
13
期号
2
页码范围
756-768
出版商
IEEE
简介
The advancement of Human-Robot Interaction (HRI) drives research into the development of advanced emotion identification architectures that fathom audio-visual (A-V) modalities of human emotion. State-of-the-art methods in multi-modal emotion recognition mainly focus on the classification of complete video sequences, leading to systems with no online potentialities. Such techniques are capable of predicting emotions only when the videos are concluded, thus restricting their applicability in practical scenarios. This article provides a novel paradigm for online emotion classification, which exploits both audio and visual modalities and produces a responsive prediction when the system is confident enough. We propose two deep Convolutional Neural Network (CNN) models for extracting emotion features, one for each modality, and a Deep Neural Network (DNN) for their fusion. In order to conceive the temporal …
引用总数
20202021202220232024411213815
学术搜索中的文章
I Kansizoglou, L Bampis, A Gasteratos - IEEE Transactions on Affective Computing, 2019