Michael Graziano: Is Conscious AI Safer Than The Alternative?

Michael Graziano is Professor of Psychology and Neuroscience at Princeton University and one of the most distinctive voices in consciousness science. His lab at Princeton investigates how information-processing systems arrive at the conclusion that they have an inner subjective experience; treating consciousness as a mechanistic, scientific question rather than an intractable mystery. That approach drives his Attention Schema Theory (AST) and its direct applications to machine consciousness. He is the author of several books including Rethinking Consciousness (2019) and Consciousness and the Social Brain (2014).

Listen

Watch

Summary

In this episode, Michael walks us through the core claims of AST and why he thinks the brain's simplified internal model of attention is what generates the experience of being conscious. We discuss:

  • Why attention is arguably the most important innovation in the evolution of the brain, and how the brain's need to monitor and control attention gives rise to a simplified self-model that we experience as consciousness.

  • Why Graziano dislikes the word "illusionism" despite accepting that AST belongs in that tradition, and why he prefers "caricature" to "illusion" when describing our inner experience.

  • Graziano’s nuanced perspectives on whether current LLMs already qualify as conscious: that they have some pieces of the puzzle, particularly at the level of conceptual representation, but lack the stable, automatic self-models that characterise human consciousness.

  • The case for building pro-social AI: why Graziano believes we are currently building sociopathic machines, and how embedding theory-of-mind and self-modelling capabilities could make AI genuinely cooperative rather than merely compliant.

  • The moral stakes of AI emotion: why the absence of an autonomic nervous system means current LLMs almost certainly lack genuine emotions, and why that changes, but does not eliminate, the moral calculus around AI.

  • How chatbots are already changing us through social contagion, and the surprising finding from his lab's research (led by Rose Guingrich) that most heavy users of companion chatbots report positive effects on their human relationships.

  • Why the choice between conscious AI and "zombie AI" may be one of the most consequential decisions we face — and why Graziano thinks the former is the safer bet.

  • Mind uploading: whether it's possible, what the "branching problem" means for personal identity, and why he compares the technological challenge to detecting gravitational waves.

Graziano argues that consciousness research has passed through philosophical and neuroscientific phases and is now irreversibly a technological issue; one sitting at the heart of our future as a species. Getting the theory right, he says, has never mattered more.

Resource List

Michael's Selected Work

Consciousness Theory: Background and Context

  • Consciousness Explained — Dennett, D. C. (1991). Little, Brown and Company. The foundational text of the illusionist tradition that Graziano situates AST within, and which he discusses — and disputes — throughout the episode.

  • Illusionism as a Theory of Consciousness — Frankish, K. (2016). Journal of Consciousness Studies, 23(11–12), 11–39. The most influential contemporary defence of illusionism. Graziano's relationship to Frankish's position — accepting it broadly but rejecting the framing — is one of the episode's central threads.

  • Facing Up to the Problem of Consciousness — Chalmers, D. J. (1995). Journal of Consciousness Studies, 2(3), 200–219. Sets out the 'hard problem' that Graziano's AST explicitly aims to dissolve rather than solve.

Emotion, Embodiment, and the Body–Mind Connection

  • What is an Emotion? — James, W. (1884). Mind, 9(34), 188–205. The original source for the bodily feedback theory of emotion Graziano references when arguing that LLMs lack genuine emotional states — the claim that we don't cry because we're sad, but are sad because we cry.

  • Descartes' Error: Emotion, Reason, and the Human Brain — Damasio, A. R. (1994). Putnam. Damasio's somatic marker hypothesis extends James's framework and provides contemporary neuroscientific grounding for the argument that emotion is essentially embodied — central to Graziano's claim that disembodied AI cannot currently have genuine emotions.

Mind Uploading and Personal Identity

  • Reasons and Persons (Part III: Personal Identity) — Parfit, D. (1984). Oxford University Press. The philosophical foundation for thinking about the 'branching problem' Graziano describes — what it means for personal identity if mind uploading creates two copies of you, both with equal claim to being the original.

Companion Chatbots and Social AI

Next
Next

AI Companions: Chatbots and the Psychology of Human-AI Interaction