Benchmarks for automated commonsense reasoning: A survey

E Davis - ACM Computing Surveys, 2023 - dl.acm.org
More than one hundred benchmarks have been developed to test the commonsense
knowledge and commonsense reasoning abilities of artificial intelligence (AI) systems …

Llm+ p: Empowering large language models with optimal planning proficiency

B Liu, Y Jiang, X Zhang, Q Liu, S Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated remarkable zero-shot generalization
abilities: state-of-the-art chatbots can provide plausible answers to many common questions …

Gpt-4v (ision) for robotics: Multimodal task planning from human demonstration

N Wake, A Kanehira, K Sasabuchi… - IEEE Robotics and …, 2024 - ieeexplore.ieee.org
We introduce a pipeline that enhances a general-purpose Vision Language Model, GPT-4V
(ision), to facilitate one-shot visual teaching for robotic manipulation. This system analyzes …

Chatgpt empowered long-step robot control in various environments: A case application

N Wake, A Kanehira, K Sasabuchi, J Takamatsu… - IEEE …, 2023 - ieeexplore.ieee.org
This paper introduces a novel method for translating natural-language instructions into
executable robot actions using OpenAI's ChatGPT in a few-shot setting. We propose …

Language-conditioned learning for robotic manipulation: A survey

H Zhou, X Yao, Y Meng, S Sun, Z BIng, K Huang… - arXiv preprint arXiv …, 2023 - arxiv.org
Language-conditioned robotic manipulation represents a cutting-edge area of research,
enabling seamless communication and cooperation between humans and robotic agents …

What foundation models can bring for robot learning in manipulation: A survey

D Li, Y Jin, H Yu, J Shi, X Hao, P Hao, H Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
The realization of universal robots is an ultimate goal of researchers. However, a key hurdle
in achieving this goal lies in the robots' ability to manipulate objects in their unstructured …

Towards open-world grasping with large vision-language models

G Tziafas, H Kasaei - arXiv preprint arXiv:2406.18722, 2024 - arxiv.org
The ability to grasp objects in-the-wild from open-ended language instructions constitutes a
fundamental challenge in robotics. An open-world grasping system should be able to …

Grounding classical task planners via vision-language models

X Zhang, Y Ding, S Amiri, H Yang, A Kaminski… - arXiv preprint arXiv …, 2023 - arxiv.org
Classical planning systems have shown great advances in utilizing rule-based human
knowledge to compute accurate plans for service robots, but they face challenges due to the …

Human demonstrations are generalizable knowledge for robots

T Cui, G Chen, T Zhou, Z Peng, M Hu, H Lu, H Li… - arXiv preprint arXiv …, 2023 - arxiv.org
Learning from human demonstrations is an emerging trend for designing intelligent robotic
systems. However, previous methods typically regard videos as instructions, simply dividing …

PlanCollabNL: leveraging Large Language Models for adaptive plan generation in human-robot collaboration

S Izquierdo-Badiola, G Canal, C Rizzo… - … on Robotics and …, 2024 - ieeexplore.ieee.org
" Hey, robot. Let's tidy up the kitchen. By the way, I have back pain today". How can a robotic
system devise a shared plan with an appropriate task allocation from this abstract goal and …