Video event detection and summarization using audio, visual and text saliency

G Evangelopoulos, A Zlatintsi… - … on acoustics, speech …, 2009 - ieeexplore.ieee.org
Detection of perceptually important video events is formulated here on the basis of saliency
models for the audio, visual and textual information conveyed in a video stream. Audio …

Divide-and-conquer based summarization framework for extracting affective video content

I Mehmood, M Sajjad, S Rho, SW Baik - Neurocomputing, 2016 - Elsevier
Recent advances in multimedia technology have led to tremendous increases in the
available volume of video data, thereby creating a major requirement for efficient systems to …

Movie summarization based on audiovisual saliency detection

G Evangelopoulos, K Rapantzikos… - 2008 15th IEEE …, 2008 - ieeexplore.ieee.org
Based on perceptual and computational attention modeling studies, we formulate measures
of saliency for an audiovisual stream. Audio saliency is captured by signal modulations and …

Real-time dual-band haptic music player for mobile devices

I Hwang, H Lee, S Choi - IEEE transactions on haptics, 2013 - ieeexplore.ieee.org
We introduce a novel dual-band haptic music player for real-time simultaneous vibrotactile
playback with music in mobile devices. Our haptic music player features a new miniature …

Audio matters in visual attention

Y Chen, TV Nguyen, M Kankanhalli… - … on Circuits and …, 2014 - ieeexplore.ieee.org
There is a dearth of information on how perceived auditory information guides image-
viewing behavior. To investigate auditory-driven visual attention, we first generated a human …

Give ear to my face: Modelling multimodal attention to social interactions

G Boccignone, V Cuculo, A D'Amelio… - Proceedings of the …, 2018 - openaccess.thecvf.com
We address the deployment of perceptual attention to social interactions as displayed in
conversational clips, when relying on multimodal information (audio and video). A …

On gaze deployment to audio-visual cues of social interactions

G Boccignone, V Cuculo, A D'Amelio, G Grossi… - IEEE …, 2020 - ieeexplore.ieee.org
Attention supports our urge to forage on social cues. Under certain circumstances, we spend
the majority of time scrutinising people, markedly their eyes and faces, and spotting persons …

Assessing the effect of physical differences in the articulation of consonants and vowels on audiovisual temporal perception

A Vatakis, P Maragos, I Rodomagoulakis… - Frontiers in integrative …, 2012 - frontiersin.org
We investigated how the physical differences associated with the articulation of speech
affect the temporal aspects of audiovisual speech perception. Video clips of consonants and …

Panoramic attention for humanoid robots

RK Sarvadevabhatla, V Ng-Thow-Hing - US Patent 8,406,925, 2013 - Google Patents
(57) ABSTRACT A robot using less storage and computational resources to embody
panoramic attention. The robot includes a panoramic attention module with multiple levels …

A biologically motivated, proto-object-based audiovisual saliency model

S Ramenahalli - AI, 2020 - mdpi.com
The natural environment and our interaction with it are essentially multisensory, where we
may deploy visual, tactile and/or auditory senses to perceive, learn and interact with our …