Multimodal local-global attention network for affective video content analysis

Y Ou, Z Chen, F Wu - … Transactions on Circuits and Systems for …, 2020 - ieeexplore.ieee.org
With the rapid development of video distribution and broadcasting, affective video content
analysis has attracted a lot of research and development activities recently. Predicting …

Affective video content analysis with adaptive fusion recurrent network

Y Yi, H Wang, Q Li - IEEE Transactions on Multimedia, 2019 - ieeexplore.ieee.org
Affective video content analysis is an important research topic in video content analysis and
has extensive applications. Intuitively, multimodal features can depict elicited emotions, and …

Knowledge-augmented multimodal deep regression bayesian networks for emotion video tagging

S Wang, L Hao, Q Ji - IEEE Transactions on Multimedia, 2019 - ieeexplore.ieee.org
The immanent dependencies between audio and visual modalities extracted from video
content and the well-established film grammar (ie, domain knowledge) are important for …

Multi-modal learning for affective content analysis in movies

Y Yi, H Wang - Multimedia Tools and Applications, 2019 - Springer
Affective content analysis is an important research topic in video content analysis, and has
extensive applications in many fields. However, it is a challenging task to design a …

Multimodal deep denoise framework for affective video content analysis

Y Zhu, Z Chen, F Wu - Proceedings of the 27th ACM international …, 2019 - dl.acm.org
Affective video content analysis has attracted a lot of attention recently. However, it faces
various challenges such as the gap between intrinsic visual-aural features and spontaneous …

A multimodal deep regression bayesian network for affective video content analyses

Q Gan, S Wang, L Hao, Q Ji - Proceedings of the IEEE …, 2017 - openaccess.thecvf.com
The inherent dependencies between visual elements and aural elements are crucial for
affective video content analyses, yet have not been successfully exploited. Therefore, we …

Violent scene detection of film videos based on multi-task learning of temporal-spatial features

Z Zheng, W Zhong, L Ye, L Fang… - 2021 IEEE 4th …, 2021 - ieeexplore.ieee.org
In this paper, we propose a new framework for the violent scene detection of film videos
based on multi-task learning of temporal-spatial features. In the proposed framework, for the …

Breaking down violence detection: Combining divide-et-impera and coarse-to-fine strategies

E Acar, F Hopfgartner, S Albayrak - Neurocomputing, 2016 - Elsevier
In today׳ s society where audio–visual content is ubiquitous, violence detection in movies
and Web videos has become a decisive functionality, eg, for providing automated youth …

Deep sentiment features of context and faces for affective video analysis

C Baecchi, T Uricchio, M Bertini… - Proceedings of the 2017 …, 2017 - dl.acm.org
Given the huge quantity of hours of video available on video sharing platforms such as
YouTube, Vimeo, etc. development of automatic tools that help users find videos that fit their …

What should we pay attention to when classifying violent videos?

MV Adão Teixeira, S Avila - … of the 16th International Conference on …, 2021 - dl.acm.org
Many works on violent video classification have proposed solutions ranging from local
descriptors to deep neural networks. Most approaches use the entire representation of the …