Contrast, Imitate, Adapt: Learning Robotic Skills From Raw Human Videos

Z Qian, M You, H Zhou, X Xu, H Fu… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Learning robotic skills from raw human videos remains a non-trivial challenge. Previous
works tackled this problem by leveraging behavior cloning or learning reward functions from …

Xskill: Cross embodiment skill discovery

M Xu, Z Xu, C Chi, M Veloso… - Conference on Robot …, 2023 - proceedings.mlr.press
Human demonstration videos are a widely available data source for robot learning and an
intuitive user interface for expressing desired behavior. However, directly extracting …

Learning video-conditioned policies for unseen manipulation tasks

E Chane-Sane, C Schmid… - 2023 IEEE International …, 2023 - ieeexplore.ieee.org
The ability to specify robot commands by a non-expert user is critical for building generalist
agents capable of solving a large variety of tasks. One convenient way to specify the …

Imitation Learning of Robotic Arm with Hierarchical Training Based on Human Videos

J Liu, J Shi, J Hou, Z Liu, W He - 2024 39th Youth Academic …, 2024 - ieeexplore.ieee.org
Human videos encapsulate a diverse spectrum of operational behaviors, serving as
invaluable pedagogical resources for imitation learning. In this paper, we introduce an …

Learning reward functions for robotic manipulation by observing humans

M Alakuijala, G Dulac-Arnold, J Mairal… - … on Robotics and …, 2023 - ieeexplore.ieee.org
Observing a human demonstrator manipulate objects provides a rich, scalable and
inexpensive source of data for learning robotic policies. However, transferring skills from …

Avid: Learning multi-stage tasks via pixel-level translation of human videos

L Smith, N Dhawan, M Zhang, P Abbeel… - arXiv preprint arXiv …, 2019 - arxiv.org
Robotic reinforcement learning (RL) holds the promise of enabling robots to learn complex
behaviors through experience. However, realizing this promise for long-horizon tasks in the …

VIEW: Visual Imitation Learning with Waypoints

A Jonnavittula, S Parekh, DP Losey - arXiv preprint arXiv:2404.17906, 2024 - arxiv.org
Robots can use Visual Imitation Learning (VIL) to learn everyday tasks from video
demonstrations. However, translating visual observations into actionable robot policies is …

Adversarial skill networks: Unsupervised robot skill learning from video

O Mees, M Merklinger, G Kalweit… - 2020 IEEE International …, 2020 - ieeexplore.ieee.org
Key challenges for the deployment of reinforcement learning (RL) agents in the real world
are the discovery, representation and reuse of skills in the absence of a reward function. To …

Giving Robots a Hand: Broadening Generalization via Hand-Centric Human Video Demonstrations

MJ Kim, J Wu, C Finn - 2022 - openreview.net
Videos of humans performing tasks are a promising data source for robotic manipulation,
because they are easy to collect in a wide range of scenarios and thus have the potential to …

Large-scale actionless video pre-training via discrete diffusion for efficient policy learning

H He, C Bai, L Pan, W Zhang, B Zhao, X Li - arXiv preprint arXiv …, 2024 - arxiv.org
Learning a generalist embodied agent capable of completing multiple tasks poses
challenges, primarily stemming from the scarcity of action-labeled robotic datasets. In …