When a human receives a prediction or recommended course of action from an intelligent agent, what additional information, beyond the prediction or recommendation itself, does the human require from the agent to decide whether to trust or reject the prediction or recommendation? In this paper we survey literature in the area of trust between a single human supervisor and a single agent subordinate to determine the nature and extent of this additional information and to characterize it into a taxonomy that can be leveraged by future researchers and intelligent agent practitioners. By examining this question from a human-centered, information-focused point of view, we can begin to compare and contrast different implementations and also provide insight and directions for future work.
arxiv.org/abs/2205.02987
- Utility Engineering: Analyzing and Controlling Emergent Value Systems in AIs - February 13, 2025
- Identifying trust cues: How trust in science is mediated in content about science - February 11, 2025
- The Importance of Intellectual Humility for Fostering Effective Communication Between Scientists and the Public - February 9, 2025