Human interaction categorization by using audio-visual cues

MJ Marín-Jiménez, R Muñoz-Salinas… - Machine vision and …, 2014 - Springer
Machine vision and applications, 2014Springer
Abstract Human Interaction Recognition (HIR) in uncontrolled TV video material is a very
challenging problem because of the huge intra-class variability of the classes (due to large
differences in the way actions are performed, lighting conditions and camera viewpoints,
amongst others) as well as the existing small inter-class variability (eg, the visual difference
between hug and kiss is very subtle). Most of previous works have been focused only on
visual information (ie, image signal), thus missing an important source of information present …
Abstract
Human Interaction Recognition (HIR) in uncontrolled TV video material is a very challenging problem because of the huge intra-class variability of the classes (due to large differences in the way actions are performed, lighting conditions and camera viewpoints, amongst others) as well as the existing small inter-class variability (e.g., the visual difference between hug and kiss is very subtle). Most of previous works have been focused only on visual information (i.e., image signal), thus missing an important source of information present in human interactions: the audio. So far, such approaches have not shown to be discriminative enough. This work proposes the use of Audio-Visual Bag of Words (AVBOW) as a more powerful mechanism to approach the HIR problem than the traditional Visual Bag of Words (VBOW). We show in this paper that the combined use of video and audio information yields to better classification results than video alone. Our approach has been validated in the challenging TVHID dataset showing that the proposed AVBOW provides statistically significant improvements over the VBOW employed in the related literature.
Springer
以上显示的是最相近的搜索结果。 查看全部搜索结果