Evaluating the Impact of Personalized Value Alignment in Human-Robot Interaction: Insights into Trust and Team Performance Outcomes

被引:5
作者
Bhat, Shreyas [1 ]
Lyons, Joseph B. [2 ]
Shi, Cong [3 ]
Yang, X. Jessie [1 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
[2] Air Force Res Lab, Dayton, OH USA
[3] Miami Herbert Business Sch, Miami, FL USA
来源
PROCEEDINGS OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024 | 2024年
关键词
Human-robot teaming; trust-aware decision-making; value-alignment; HUMAN-MACHINE COLLABORATION; TRANSPARENCY-BASED FEEDBACK; DRIVERS TRUST;
D O I
10.1145/3610977.3634921
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper examines the effect of real-time, personalized alignment of a robot's reward function to the human's values on trust and team performance. We present and compare three distinct robot interaction strategies: a non-learner strategy where the robot presumes the human's reward function mirrors its own; a non-adaptive-learner strategy in which the robot learns the human's reward function for trust estimation and human behavior modeling, but still optimizes its own reward function; and an adaptive-learner strategy in which the robot learns the human's reward function and adopts it as its own. Two human-subject experiments with a total number of N = 54 participants were conducted. In both experiments, the human-robot team searches for potential threats in a town. The team sequentially goes through search sites to look for threats. We model the interaction between the human and the robot as a trust-aware Markov Decision Process (trust-aware MDP) and use Bayesian Inverse Reinforcement Learning (IRL) to estimate the reward weights of the human as they interact with the robot. In Experiment 1, we start our learning algorithm with an informed prior of the human's values/goals. In Experiment 2, we start the learning algorithm with an uninformed prior. Results indicate that when starting with a good informed prior, personalized value alignment does not seem to benefit trust or team performance. On the other hand, when an informed prior is unavailable, alignment to the human's values leads to high trust and higher perceived performance while maintaining the same objective team performance.
引用
收藏
页码:32 / 41
页数:10
相关论文
共 45 条
  • [1] Improving Human-Machine Collaboration Through Transparency-based Feedback - Part I: Human Trust and Workload Model
    Akash, Kumar
    Polson, Katelyn
    Reid, Tahira
    Jain, Neera
    [J]. IFAC PAPERSONLINE, 2019, 51 (34): : 315 - 321
  • [2] Improving Human-Machine Collaboration Through Transparency-based Feedback - Part II: Control Design and Synthesis
    Akash, Kumar
    Reid, Tahira
    Jain, Neera
    [J]. IFAC PAPERSONLINE, 2019, 51 (34): : 322 - 328
  • [3] A survey of inverse reinforcement learning: Challenges, methods and progress
    Arora, Saurabh
    Doshi, Prashant
    [J]. ARTIFICIAL INTELLIGENCE, 2021, 297 (297)
  • [4] Context-Adaptive Management of Drivers' Trust in Automated Vehicles
    Azevedo-Sa, Hebert
    Jayaraman, Suresh Kumaar
    Yang, X. Jessie
    Robert, Lionel P., Jr.
    Tilbury, Dawn M.
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6908 - 6915
  • [5] Clustering Trust Dynamics in a Human-Robot Sequential Decision-Making Task
    Bhat, Shreyas
    Lyons, Joseph B.
    Shi, Cong
    Yang, X. Jessie
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04): : 8815 - 8822
  • [6] Billings DR, 2012, ACMIEEE INT CONF HUM, P109
  • [7] Biyik E, 2018, Arxiv, DOI arXiv:1810.04303
  • [8] Trust-Aware Decision Making for Human-Robot Collaboration: Model Learning and Planning
    Chen, Min
    Nikolaidis, Stefanos
    Soh, Harold
    Hsu, David
    Srinivasa, Siddhartha
    [J]. ACM TRANSACTIONS ON HUMAN-ROBOT INTERACTION, 2020, 9 (02)
  • [9] Planning with Trust for Human-Robot Collaboration
    Chen, Min
    Nikolaidis, Stefanos
    Soh, Harold
    Hsu, David
    Srinivasa, Siddhartha
    [J]. HRI '18: PROCEEDINGS OF THE 2018 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2018, : 307 - 315
  • [10] Chiou Erin K., 2023, Trusting Automation: Designing for Responsivity and Resilience, V65, P137, DOI 10.HumanFactors1177/00187208211009995