Mobile aloha: Learning bimanual mobile manipulation with low-cost whole-body teleoperation

Z Fu, TZ Zhao, C Finn - arXiv preprint arXiv:2401.02117, 2024 - arxiv.org
Imitation learning from human demonstrations has shown impressive performance in
robotics. However, most results focus on table-top manipulation, lacking the mobility and …

Masked world models for visual control

Y Seo, D Hafner, H Liu, F Liu, S James… - … on Robot Learning, 2023 - proceedings.mlr.press
Visual model-based reinforcement learning (RL) has the potential to enable sample-efficient
robot learning from visual observations. Yet the current approaches typically train a single …

Mimicplay: Long-horizon imitation learning by watching human play

C Wang, L Fan, J Sun, R Zhang, L Fei-Fei, D Xu… - arXiv preprint arXiv …, 2023 - arxiv.org
Imitation learning from human demonstrations is a promising paradigm for teaching robots
manipulation skills in the real world. However, learning complex long-horizon tasks often …

Videodex: Learning dexterity from internet videos

K Shaw, S Bahl, D Pathak - Conference on Robot Learning, 2023 - proceedings.mlr.press
To build general robotic agents that can operate in many environments, it is often imperative
for the robot to collect experience in the real world. However, this is often not feasible due to …

Robotap: Tracking arbitrary points for few-shot visual imitation

M Vecerik, C Doersch, Y Yang… - … on Robotics and …, 2024 - ieeexplore.ieee.org
For robots to be useful outside labs and specialized factories we need a way to teach them
new useful behaviors quickly. Current approaches lack either the generality to onboard new …

Concept2robot: Learning manipulation concepts from instructions and human demonstrations

L Shao, T Migimatsu, Q Zhang… - … Journal of Robotics …, 2021 - journals.sagepub.com
We aim to endow a robot with the ability to learn manipulation concepts that link natural
language instructions to motor skills. Our goal is to learn a single multi-task policy that takes …

Meta-reward-net: Implicitly differentiable reward learning for preference-based reinforcement learning

R Liu, F Bai, Y Du, Y Yang - Advances in Neural …, 2022 - proceedings.neurips.cc
Abstract Setting up a well-designed reward function has been challenging for many
reinforcement learning applications. Preference-based reinforcement learning (PbRL) …

Diff-lfd: Contact-aware model-based learning from visual demonstration for robotic manipulation via differentiable physics-based simulation and rendering

X Zhu, JH Ke, Z Xu, Z Sun, B Bai, J Lv… - … on Robot Learning, 2023 - proceedings.mlr.press
Abstract Learning from Demonstration (LfD) is an efficient technique for robots to acquire
new skills through expert observation, significantly mitigating the need for laborious manual …

Towards a unified agent with foundation models

N Di Palo, A Byravan, L Hasenclever… - arXiv preprint arXiv …, 2023 - arxiv.org
Language Models and Vision Language Models have recently demonstrated
unprecedented capabilities in terms of understanding human intentions, reasoning, scene …

Imitation learning: Progress, taxonomies and challenges

B Zheng, S Verma, J Zhou, IW Tsang… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Imitation learning (IL) aims to extract knowledge from human experts' demonstrations or
artificially created agents to replicate their behaviors. It promotes interdisciplinary …