This work proposes a framework that incorporates trust in an ad hoc teamwork scenario with human-agent teams, where an agent must collaborate with a human to perform a task. During the task, the agent must infer, through interactions and observations, how much the human trusts it and adapt its behaviour to maximize the team’s performance. To achieve this, we propose collecting data from human participants in experiments to define different settings (based on trust levels) and learning optimal policies for each of them. Then, we create a module to infer the current setting (depending on the amount of trust). Finally, we validate this framework in a real-world scenario and analyse how this adaptable behaviour affects trust.
Latest posts by Ryan Watkins (see all)
- Using Digital Nudges To Enhance Collective Intelligence In Online Collaboration: Insights From Unexpected Outcomes. - March 15, 2024
- Enhancing Instructional Quality: Leveraging Computer-Assisted Textual Analysis to Generate In-Depth Insights from Educational Artifacts - March 7, 2024
- Attraction effect in crowded decision spaces: exploring the impact of decoys in choices among numerous options - March 2, 2024