Rvt: Robotic view transformer for 3d object manipulation

A Goyal, J Xu, Y Guo, V Blukis… - Conference on Robot …, 2023 - proceedings.mlr.press
For 3D object manipulation, methods that build an explicit 3D representation perform better
than those relying only on camera images. But using explicit 3D representations like voxels …

Fastrlap: A system for learning high-speed driving via deep rl and autonomous practicing

K Stachowicz, D Shah, A Bhorkar… - … on Robot Learning, 2023 - proceedings.mlr.press
We present a system that enables an autonomous small-scale RC car to drive aggressively
from visual observations using reinforcement learning (RL). Our system, FastRLAP, trains …

One-4-all: Neural potential fields for embodied navigation

S Morin, M Saavedra-Ruiz… - 2023 IEEE/RSJ …, 2023 - ieeexplore.ieee.org
A fundamental task in robotics is to navigate between two locations. In particular, real-world
navigation can require long-horizon planning using high-dimensional RGB images, which …

Fastrlap: A system for learning high-speed driving via deep rl and autonomous practicing

D Shah, K Stachowicz, A Bhorkar… - … on Pretraining for …, 2023 - openreview.net
We present a system that enables an autonomous small-scale rally car to drive at high
speed from only image observations using reinforcement learning. Our method trains …

Visual Semantic Navigation with Real Robots

C Gutiérrez-Álvarez, P Ríos-Navarro… - arXiv preprint arXiv …, 2023 - arxiv.org
Visual Semantic Navigation (VSN) is the ability of a robot to learn visual semantic
information for navigating in unseen environments. These VSN models are typically tested in …

Learning Speed Adaptation for Flight in Clutter

G Zhao, T Wu, Y Chen, F Gao - IEEE Robotics and Automation …, 2024 - ieeexplore.ieee.org
Animals learn to adapt speed of their movements to their capabilities and the environment
they observe. Mobile robots should also demonstrate this ability to trade-off aggressiveness …

Learning Agility Adaptation for Flight in Clutter

G Zhao, T Wu, Y Chen, F Gao - arXiv preprint arXiv:2403.04586, 2024 - arxiv.org
Animals learn to adapt agility of their movements to their capabilities and the environment
they operate in. Mobile robots should also demonstrate this ability to combine agility and …

Empowering Embodied Visual Tracking with Visual Foundation Models and Offline RL

F Zhong, K Wu, H Ci, C Wang, H Chen - arXiv preprint arXiv:2404.09857, 2024 - arxiv.org
Embodied visual tracking is to follow a target object in dynamic 3D environments using an
agent's egocentric vision. This is a vital and challenging skill for embodied agents. However …

SCALE: Self-Correcting Visual Navigation for Mobile Robots via Anti-Novelty Estimation

C Chen, Y Liu, Y Zhuang, S Mao, K Xue… - arXiv preprint arXiv …, 2024 - arxiv.org
Although visual navigation has been extensively studied using deep reinforcement learning,
online learning for real-world robots remains a challenging task. Recent work directly …

Risk-Sensitive Mobile Robot Navigation in Crowded Environment via Offline Reinforcement Learning

J Wu, Y Wang, H Asama, Q An… - 2023 IEEE/RSJ …, 2023 - ieeexplore.ieee.org
Mobile robot navigation in a human-populated environment has been of great interest to the
research community in recent years, referred to as crowd navigation. Currently, offline …