Title: Probing and Persuading Moral and Metaphysical Beliefs with Human–AI Dialogues
Abstract: Josh will present two projects using human–AI dialogues to study and influence moral and metaphysical beliefs. The first tests whether conversations with LLMs can shift belief in God, challenging the idea that religious convictions are uniquely resistant to reasoned persuasion. The second revisits Jonathan Haidt’s “moral dumbfounding” hypothesis—the idea that people often judge acts as wrong even when they cannot defend that judgment with harm-based reasons—by using adaptive AI dialogues to argue that taboo actions (e.g., consensual incest, cannibalism) depicted in vignettes cause no harm, and observing how this shapes subsequent moral judgments. This project brings an oft-cited but weakly evidenced social psychological phenomenon to its strongest test yet, harnessing the persuasive power of modern AI.