We introduce BridgeData V2, a large and diverse dataset of robotic manipulation behaviors designed to facilitate research in scalable robot learning. BridgeData V2 contains 53,896 …
Massive datasets and high-capacity models have driven many recent advancements in computer vision and natural language understanding. This work presents a platform to …
We present BEHAVIOR-1K, a comprehensive simulation benchmark for human-centered robotics. BEHAVIOR-1K includes two components, guided and motivated by the results of an …
Abstract 3D vision-language (3D-VL) grounding, which aims to align language with 3D physical environments, stands as a cornerstone in developing embodied agents. In …
Y Xu, W Wan, J Zhang, H Liu, Z Shan… - Proceedings of the …, 2023 - openaccess.thecvf.com
In this work, we tackle the problem of learning universal robotic dexterous grasping from a point cloud observation under a table-top setting. The goal is to grasp and lift up objects in …
Y Liu, Y Liu, C Jiang, K Lyu, W Wan… - Proceedings of the …, 2022 - openaccess.thecvf.com
We present HOI4D, a large-scale 4D egocentric dataset with rich annotations, to catalyze the research of category-level human-object interaction. HOI4D consists of 2.4 M RGB-D …
We propose a novel, object-agnostic method for learning a universal policy for dexterous object grasping from realistic point cloud observations and proprioceptive information under …
R Yang, G Yang, X Wang - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
Legged robots have the potential to expand the reach of autonomy beyond paved roads. In this work, we consider the difficult problem of locomotion on challenging terrains using a …
For years, researchers have been devoted to generalizable object perception and manipulation, where cross-category generalizability is highly desired yet underexplored. In …