Human activity recognition in artificial intelligence framework: a narrative review

N Gupta, SK Gupta, RK Pathak, V Jain… - Artificial intelligence …, 2022 - Springer
Human activity recognition (HAR) has multifaceted applications due to its worldly usage of
acquisition devices such as smartphones, video cameras, and its ability to capture human …

A survey of robot learning strategies for human-robot collaboration in industrial settings

D Mukherjee, K Gupta, LH Chang, H Najjaran - Robotics and Computer …, 2022 - Elsevier
Increased global competition has placed a premium on customer satisfaction, and there is a
greater demand for manufacturers to be flexible with their products and services. This …

Ego4d: Around the world in 3,000 hours of egocentric video

K Grauman, A Westbury, E Byrne… - Proceedings of the …, 2022 - openaccess.thecvf.com
We introduce Ego4D, a massive-scale egocentric video dataset and benchmark suite. It
offers 3,670 hours of daily-life activity video spanning hundreds of scenarios (household …

Affordances from human videos as a versatile representation for robotics

S Bahl, R Mendonca, L Chen… - Proceedings of the …, 2023 - openaccess.thecvf.com
Building a robot that can understand and learn to interact by watching humans has inspired
several vision problems. However, despite some successful results on static datasets, it …

Anticipative video transformer

R Girdhar, K Grauman - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Abstract We propose Anticipative Video Transformer (AVT), an end-to-end attention-based
video modeling architecture that attends to the previously observed video in order to …

Vision-based holistic scene understanding towards proactive human–robot collaboration

J Fan, P Zheng, S Li - Robotics and Computer-Integrated Manufacturing, 2022 - Elsevier
Recently human–robot collaboration (HRC) has emerged as a promising paradigm for mass
personalization in manufacturing owing to the potential to fully exploit the strength of human …

Rescaling egocentric vision: Collection, pipeline and challenges for epic-kitchens-100

D Damen, H Doughty, GM Farinella, A Furnari… - International Journal of …, 2022 - Springer
This paper introduces the pipeline to extend the largest dataset in egocentric vision, EPIC-
KITCHENS. The effort culminates in EPIC-KITCHENS-100, a collection of 100 hours, 20M …

Vln bert: A recurrent vision-and-language bert for navigation

Y Hong, Q Wu, Y Qi… - Proceedings of the …, 2021 - openaccess.thecvf.com
Accuracy of many visiolinguistic tasks has benefited significantly from the application of
vision-and-language (V&L) BERT. However, its application for the task of vision-and …

Dynamic multiscale graph neural networks for 3d skeleton based human motion prediction

M Li, S Chen, Y Zhao, Y Zhang… - Proceedings of the …, 2020 - openaccess.thecvf.com
We propose novel dynamic multiscale graph neural networks (DMGNN) to predict 3D
skeleton-based human motions. The core idea of DMGNN is to use a multiscale graph to …

Human action recognition and prediction: A survey

Y Kong, Y Fu - International Journal of Computer Vision, 2022 - Springer
Derived from rapid advances in computer vision and machine learning, video analysis tasks
have been moving from inferring the present state to predicting the future state. Vision-based …