Learning by Watching: A Review of Video-based Learning Approaches for Robot Manipulation

C Eze, C Crick - arXiv preprint arXiv:2402.07127, 2024 - arxiv.org
Robot learning of manipulation skills is hindered by the scarcity of diverse, unbiased
datasets. While curated datasets can help, challenges remain in generalizability and real …

[HTML][HTML] A survey of demonstration learning

A Correia, LA Alexandre - Robotics and Autonomous Systems, 2024 - Elsevier
With the fast improvement of machine learning, reinforcement learning (RL) has been used
to automate human tasks in different areas. However, training such agents is difficult and …

Attentive one-shot meta-imitation learning from visual demonstration

V Bhutani, A Majumder, M Vankadari… - … on Robotics and …, 2022 - ieeexplore.ieee.org
The ability to apply a previously-learned skill (eg, pushing) to a new task (context or object)
is an important requirement for new-age robots. An attempt is made to solve this problem in …

Contrastively learning visual attention as affordance cues from demonstrations for robotic grasping

Y Zha, S Bhambri, L Guan - 2021 IEEE/RSJ International …, 2021 - ieeexplore.ieee.org
Conventional works that learn grasping affordance from demonstrations need to explicitly
predict grasping configurations, such as gripper approaching angles or grasping preshapes …

Perception Stitching: Zero-Shot Perception Encoder Transfer for Visuomotor Robot Policies

P Jian, E Lee, Z Bell, MM Zavlanos, B Chen - arXiv preprint arXiv …, 2024 - arxiv.org
Vision-based imitation learning has shown promising capabilities of endowing robots with
various motion skills given visual observation. However, current visuomotor policies fail to …

Fov-net: Field-of-view extrapolation using self-attention and uncertainty

L Ma, S Georgoulis, X Jia… - IEEE Robotics and …, 2021 - ieeexplore.ieee.org
The ability to make educated predictions about their surroundings, and associate them with
certain confidence, is important for intelligent systems, like autonomous vehicles and robots …

Multi-view contrastive learning from demonstrations

A Correia, LA Alexandre - 2022 Sixth IEEE International …, 2022 - ieeexplore.ieee.org
This paper presents a framework for learning visual representations from unlabeled video
demonstrations captured from multiple viewpoints. We show that these representations are …

Contrastive Learning from Demonstrations

A Correia, LA Alexandre - arXiv preprint arXiv:2201.12813, 2022 - arxiv.org
This paper presents a framework for learning visual representations from unlabeled video
demonstrations captured from multiple viewpoints. We show that these representations are …

Perceiving, Planning, Acting, and Self-Explaining: A Cognitive Quartet with Four Neural Networks

Y Zha - 2022 - search.proquest.com
Learning to accomplish complex tasks may require a tight coupling among different levels of
cognitive functions or components, like perception, acting, planning, and self-explaining …

Understanding Manipulation Contexts by Vision and Language for Robotic Vision

C Jiang - 2021 - era.library.ualberta.ca
Abstract In Activities of Daily Living (ADLs), humans perform thousands of arm and hand
object manipulation tasks, such as picking, pouring and drinking a drink. Interpreting such …