Review and challenges of technologies for real-time human behavior monitoring

S Dávila-Montero, JA Dana-Lê, G Bente… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
A person's behavior significantly influences their health and well-being. It also contributes to
the social environment in which humans interact, with cascading impacts to the health and …

Coupled multimodal emotional feature analysis based on broad-deep fusion networks in human–robot interaction

L Chen, M Li, M Wu, W Pedrycz… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
A coupled multimodal emotional feature analysis (CMEFA) method based on broad–deep
fusion networks, which divide multimodal emotion recognition into two layers, is proposed …

Multi-modal anomaly detection by using audio and visual cues

AU Rehman, HS Ullah, H Farooq, MS Khan… - IEEE …, 2021 - ieeexplore.ieee.org
This paper considers the problem of anomaly detection in an outdoor environment where
surveillance cameras are usually installed to monitor activities of general public. A novel …

Speech Emotion Recognition Based on Temporal-Spatial Learnable Graph Convolutional Neural Network

J Yan, H Li, F Xu, X Zhou, Y Liu, Y Yang - Electronics, 2024 - mdpi.com
The Graph Convolutional Neural Networks (GCN) method has shown excellent performance
in the field of deep learning, and using graphs to represent speech data is a computationally …

[PDF][PDF] Mining multimodal repositories for speech affecting diseases

J Correia12, B Raj, I Trancoso, F Teixeira - 2018 - researchgate.net
The motivation for this work is to contribute to the collection of large in-the-wild multimodal
datasets in which the speech of the subject is affected by certain medical conditions. Our …

Addressing data scarcity in multimodal user state recognition by combining semi-supervised and supervised learning

H Voß, H Wersing, S Kopp - … of the 2021 International Conference on …, 2021 - dl.acm.org
Detecting mental states of human users is crucial for the development of cooperative and
intelligent robots, as it enables the robot to understand the user's intentions and desires …

Combining key pronunciation detection, frontal lip reconstruction, and time-delay for audio-visual consistency judgment

Z Zhu, C Luo, L Liao, P Lin, Y Li - Digital Signal Processing, 2024 - Elsevier
This paper presents a novel approach to audio-visual consistency judgment (AVCJ), in
which vowel-like regions exhibiting significant changes in lip shape are used as key …

[图书][B] Real-Time Human/Group Interaction Monitoring Platform Integrating Sensor Fusion and Machine Learning Approaches

S Dávila-Montero - 2022 - search.proquest.com
A person's social intelligence impacts their physical and mental health, and the productivity
levels of the individuals involved, for example, in workplace interactions. To promote …

Human activity recognition using robust adaptive privileged probabilistic learning

M Vrigkas, E Kazakos, C Nikou… - Pattern Analysis and …, 2021 - Springer
In this work, a supervised probabilistic approach is proposed that integrates the learning
using privileged information (LUPI) paradigm into a hidden conditional random field (HCRF) …

Robust incremental hidden conditional random fields for human action recognition

M Vrigkas, E Mastora, C Nikou… - Advances in Visual …, 2018 - Springer
Hidden conditional random fields (HCRFs) are a powerful supervised classification system,
which is able to capture the intrinsic motion patterns of a human action. However, finding the …