•  
  •  
 

Joint Force Quarterly

Abstract

This article explores a first-of-its-kind tabletop exercise designed to expose U.S. military and national security professionals to the real-world challenges of using AI tools like deepfake detectors in high-stakes decisionmaking. Set in a fictional scenario involving a flawed AI system, participants confronted algorithmic bias, technical uncertainty, and ethical dilemmas, revealing how AI limitations can erode trust in intelligence products. Drawing from student and faculty reflections, the article highlights the importance of preparing future leaders to scrutinize and manage AI performance failures not only as technical bugs but as complex, systemic issues with social and operational consequences. The exercise underscored the need for a more holistic, human-centered approach to AI in defense contexts.

Share

COinS