Current AI evaluation paradigms that rely on static, model-only tests fail to capture harms that emerge through sustained human-AI interaction. As interactive AI systems, such as AI companions, proliferate in daily life, this mismatch between evaluation methods and real-world use becomes increasingly consequential. We argue for a paradigm shift toward evaluation centered on \textit{interactional ethics}, which addresses risks like inappropriate human-AI relationships, social manipulation, and cognitive overreliance that develop through repeated interaction rather than single outputs. Drawing on human-computer interaction, natural language processing, and the social sciences, we propose principles for evaluating generative models through interaction scenarios and human impact metrics. We conclude by examining implementation challenges and open research questions for researchers, practitioners, and regulators integrating these approaches into AI governance frameworks.
Latest posts by Ryan Watkins (see all)
- Towards interactive evaluations for interaction harms in human-AI systems - April 29, 2025
- How Generative AI Fixes What Higher Education Broke - April 28, 2025
- AI for Accessible Education: Personalized Audio-Based Learning for Blind Students - April 25, 2025