Videodex: Learning dexterity from internet videos

K Shaw, S Bahl, D Pathak - Conference on Robot Learning, 2023 - proceedings.mlr.press
To build general robotic agents that can operate in many environments, it is often imperative
for the robot to collect experience in the real world. However, this is often not feasible due to …

Learning generalizable robotic reward functions from" in-the-wild" human videos

AS Chen, S Nair, C Finn - arXiv preprint arXiv:2103.16817, 2021 - arxiv.org
We are motivated by the goal of generalist robots that can complete a wide range of tasks
across many environments. Critical to this is the robot's ability to acquire some metric of task …

Avid: Learning multi-stage tasks via pixel-level translation of human videos

L Smith, N Dhawan, M Zhang, P Abbeel… - arXiv preprint arXiv …, 2019 - arxiv.org
Robotic reinforcement learning (RL) holds the promise of enabling robots to learn complex
behaviors through experience. However, realizing this promise for long-horizon tasks in the …

Reinforcement learning with videos: Combining offline observations with interaction

K Schmeckpeper, O Rybkin, K Daniilidis… - arXiv preprint arXiv …, 2020 - arxiv.org
Reinforcement learning is a powerful framework for robots to acquire skills from experience,
but often requires a substantial amount of online data collection. As a result, it is difficult to …

Robonet: Large-scale multi-robot learning

S Dasari, F Ebert, S Tian, S Nair, B Bucher… - arXiv preprint arXiv …, 2019 - arxiv.org
Robot learning has emerged as a promising tool for taming the complexity and diversity of
the real world. Methods based on high-capacity models, such as deep networks, hold the …

Affordances from human videos as a versatile representation for robotics

S Bahl, R Mendonca, L Chen… - Proceedings of the …, 2023 - openaccess.thecvf.com
Building a robot that can understand and learn to interact by watching humans has inspired
several vision problems. However, despite some successful results on static datasets, it …

R3m: A universal visual representation for robot manipulation

S Nair, A Rajeswaran, V Kumar, C Finn… - arXiv preprint arXiv …, 2022 - arxiv.org
We study how visual representations pre-trained on diverse human video data can enable
data-efficient learning of downstream robotic manipulation tasks. Concretely, we pre-train a …

Robotic offline rl from internet videos via value-function pre-training

C Bhateja, D Guo, D Ghosh, A Singh, M Tomar… - arXiv preprint arXiv …, 2023 - arxiv.org
Pre-training on Internet data has proven to be a key ingredient for broad generalization in
many modern ML systems. What would it take to enable such capabilities in robotic …

Graph inverse reinforcement learning from diverse videos

S Kumar, J Zamora, N Hansen… - … on Robot Learning, 2023 - proceedings.mlr.press
Abstract Research on Inverse Reinforcement Learning (IRL) from third-person videos has
shown encouraging results on removing the need for manual reward design for robotic …

Zero-shot robot manipulation from passive human videos

H Bharadhwaj, A Gupta, S Tulsiani, V Kumar - arXiv preprint arXiv …, 2023 - arxiv.org
Can we learn robot manipulation for everyday tasks, only by watching videos of humans
doing arbitrary tasks in different unstructured settings? Unlike widely adopted strategies of …