This paper presents ShapeLLM, the first 3D Multimodal Large Language Model (LLM) designed for embodied interaction, exploring a universal 3D object understanding with 3D …
We present PartNet: a consistent, large-scale dataset of 3D objects annotated with fine- grained, instance-level, and hierarchical 3D part information. Our dataset consists of …
S Deng, X Xu, C Wu, K Chen… - proceedings of the IEEE …, 2021 - openaccess.thecvf.com
The ability to understand the ways to interact with objects from visual cues, aka visual affordance, is essential to vision-guided robotic research. This involves categorizing …
This report surveys advances in deep learning‐based modelling techniques that address four different 3D indoor scene analysis tasks, as well as synthesis of 3D indoor scenes. We …
A critical aspect of human visual perception is the ability to parse visual scenes into individual objects and further into object parts, forming part-whole hierarchies. Such …
X Wang, B Zhou, Y Shi, X Chen… - Proceedings of the …, 2019 - openaccess.thecvf.com
For the task of mobility analysis of 3D shapes, we propose joint analysis for simultaneous motion part segmentation and motion attribute estimation, taking a single 3D model as input …
Data-driven methods serve an increasingly important role in discovering geometric, structural, and semantic relationships between shapes. In contrast to traditional approaches …
L Yoon, D Yang, J Kim, CH Chung… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
Rapidly developing technologies are realizing a 3D telepresence, in which geographically separated users can interact with each other through their virtual avatars. In this article, we …
We introduce a framework for action-driven evolution of 3D indoor scenes, where the goal is to simulate how scenes are altered by human actions, and specifically, by object placements …