학술논문

Examining Profiles for Robotic Risk Assessment: Does a Robot’s Approach to Risk Affect User Trust?
Document Type
Conference
Source
2020 15th ACM/IEEE International Conference on Human-Robot Interaction (HRI) Human-Robot Interaction (HRI), 2020 15th ACM/IEEE International Conference on. :23-31 Mar, 2020
Subject
Computing and Processing
Robotics and Control Systems
Knowledge engineering
Atmospheric measurements
Sociology
Human-robot interaction
Particle measurements
Reliability engineering
Risk management
Risk
Nuclear
Decision making
Trust
Performance
HRI
Prospect theory
Language
ISSN
2167-2148
Abstract
As autonomous robots move towards ubiquity, the need for robots to make decisions under risk that are trustworthy becomes increasingly significant; both to aid acceptance and to fully utilise their autonomous capabilities. We propose that incorporating a human approach to risk assessment into a robot’s decision making process will increase user trust. This work investigates four robotic approaches to risk and, through a user study, explores the levels of trust placed in each. These approaches are: risk averse, risk seeking, risk neutral and a human approach to risk. Risk is artificially stimulated through performance-based compensation, in line with previous studies. The study was conducted in a virtual nuclear environment created using the Unity games engine. Forty participants were asked to complete a robot supervision task, in which they observed a robot making risk based decisions and were able to question the robot, question the robot further and ultimately accept or alter the robot’s decision. It is shown that a robot that is risk seeking is significantly less trusted than a risk averse robot, a risk neutral robot and a robot utilising human approach to risk. There was found to be no significant difference between the levels of trust placed in the risk averse, risk neutral and human approach to risk. It is also found that the level to which participants question a robot’s decisions does not form an accurate measure of trust. The results suggest that when designing a robot that must make risk based decisions during teleoperation in a hazardous environment, an engineer should avoid a risk seeking robot. However, that same engineer may choose whichever of the remaining risk profiles best suits the implementation, with knowledge that the trust in their system is unlikely to be significantly affected. ACM Reference Format: Tom Bridgwater, Manuel Giuliani, Anouk van Maris, Greg Baker, Alan Winfield, and Tony Pipe. 2020. Examining Profiles for Robotic Risk Assessment: Does a Robot’s Approach to Risk Affect User Trust?. In Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction (HRI’20), March 23–26, 2020, Cambridge, United Kingdom. ACM, New York, NY, USA, 9 pages. https://doi.org/10.1145/3319502.3374804