作者
Shreyas Bhat, Joseph B Lyons, Cong Shi, X Jessie Yang
发表日期
2024/3
研讨会论文
Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction
页码范围
32-41
简介
This paper examines the effect of real-time, personalized alignment of a robot's reward function to the human's values on trust and team performance. We present and compare three distinct robot interaction strategies: a non-learner strategy where the robot presumes the human's reward function mirrors its own; a non-adaptive-learner strategy in which the robot learns the human's reward function for trust estimation and human behavior modeling, but still optimizes its own reward function; and an adaptive-learner strategy in which the robot learns the human's reward function and adopts it as its own. Two human-subject experiments with a total number of N=54 participants were conducted. In both experiments, the human-robot team searches for potential threats in a town. The team sequentially goes through search sites to look for threats. We model the interaction between the human and the robot as a trust-aware …
引用总数
学术搜索中的文章