In 1982, the film Blade Runner presented a world where artificial beings, called replicants, were virtually indistinguishable from humans. The film’s central tension revolves around a profound question: How can we tell if an artificial mind is truly conscious? The replicants in the film display emotion, reasoning, and even empathy, yet they’re dismissed as mere machines – much like how we might view today’s artificial intelligence systems. When one replicant, facing his final moments, says “I’ve seen things you people wouldn’t believe,” we’re forced to confront the possibility that these artificial beings might have genuine inner experiences, real consciousness.

This question has moved from science fiction into reality. As we interact with increasingly sophisticated large language models (LLMs), we face similar challenges: How can we tell if these systems possess genuine consciousness or merely convincingly simulate it?
What makes these questions particularly complex is how we naturally interpret AI behavior. Philosopher Daniel Dennett describes what he calls the ‘intentional stance‘ – our tendency to treat complex systems as if they have beliefs, desires, and intentions. When an AI assistant gives a helpful response, we might say, “it understood what I needed,” rather than considering the pattern matching that produced the output. This isn’t just loose talk – it’s a fundamental strategy humans use to make sense of complex behavior, whether from other humans, animals, or artificial systems.
In a recent analysis, philosopher David Chalmers addressed this question head-on, reminding us of what he famously called “the hard problem of consciousness”—why do physical processes give rise to subjective, first-person experiences? Even if we fully understood how information flows through artificial networks, we still wouldn’t have explained why these processes should be accompanied by conscious experience. This challenge becomes particularly relevant as we try to assess consciousness in artificial systems.
Theory of Mind and Machine Consciousness
To evaluate whether an AI system might be conscious, we need to examine a crucial capability: Theory of Mind (ToM) – our ability to understand and model others’ mental states. This capacity goes beyond simple behavioral prediction; it involves understanding that others have beliefs, desires, and intentions different from our own. In Blade Runner, the Voight-Kampff test examines this capability by presenting emotionally provocative scenarios, like asking how the subject would react to finding a tortoise struggling in the desert. The test doesn’t just look for appropriate responses; it searches for genuine empathy and understanding of others’ suffering. This approach differs significantly from Alan Turing’s original test, which focused purely on behavioral imitation through conversation. While the Turing test remains influential, its emphasis on behavior alone may not be sufficient to distinguish genuine consciousness from sophisticated simulation.
Recent empirical research has provided compelling evidence about LLMs’ Theory of Mind capabilities by directly comparing 11 models against children aged 7-10 on classic tasks. The findings reveal a striking pattern: while most LLMs perform below children’s level, the largest instruction-tuned models can sometimes surpass child performance. However, even advanced models struggle particularly with second-order theory of mind (understanding what someone thinks about another person’s thoughts), especially when researchers modify the classic test scenarios. This suggests LLMs might rely more on pattern recognition from common scenarios than developing a genuine understanding of mental states – a limitation that Emily Bender and her colleagues powerfully captured in their characterization of LLMs as “stochastic parrots,” systems that convincingly imitate language without true understanding.
Interestingly, LLMs perform notably better on tests involving non-literal language use, such as understanding irony, sarcasm, and white lies, likely because their training data includes abundant examples of such linguistic phenomena. Perhaps most intriguingly, the study points to the role of instruction tuning in developing these capabilities: just as humans develop Theory of Mind through social interaction and receiving feedback on their communication, LLMs develop better perspective-taking abilities through instruction tuning that rewards cooperative communication patterns. This parallel between human development and AI training opens up fascinating questions about the nature of social intelligence and how it emerges in different types of systems.
Cheryl’s Birthday Experiment
While emotional tests like the Voight-Kampff probe one aspect of consciousness, researchers have developed more formal ways to evaluate an AI system’s ability to understand and reason about other minds. Peter Norvig’s experiment with Cheryl’s Birthday puzzle offers a particularly revealing test case. Like the Voight-Kampff test in Blade Runner, this puzzle examines a key aspect of consciousness – but instead of testing emotional responses, it probes the ability to reason about others’ mental states precisely and logically.
The puzzle, designed initially as a logic puzzle and brain teaser that challenged human reasoning capabilities, presents a scenario where Cheryl tells Albert and Bernard ten possible dates for her birthday. She then privately tells Albert only the month and Bernard only the day. Through a series of statements, they deduce the correct date by reasoning about each other’s knowledge:
First, Albert says: “I don’t know Cheryl’s birthday, but I also know Bernard doesn’t know.” This statement reveals complex reasoning – Albert must consider what Bernard knows (just the day number) and realize that if Cheryl’s birthday were on a certain date, Bernard would have immediately known the answer. Albert’s confidence that Bernard doesn’t know tells us something important about which dates could be possible.
Then Bernard responds, “At first, I didn’t know Cheryl’s birthday, but now I do.” This shows Bernard using Albert’s statement to eliminate possibilities and reach certainty. Bernard has considered Albert’s understanding of his (Bernard’s) knowledge to update his understanding.
Finally, Albert concludes: “Now I also know Cheryl’s birthday.” The solution emerges through this dance of mutual understanding – each person uses their knowledge of what the other knows and doesn’t know to narrow down the possibilities progressively.

When Norvig tested nine leading LLMs on this puzzle, the results revealed these systems’ limitations. While the LLMs could recall the puzzle and even state its solution (July 16), this wasn’t because they understood the logical reasoning – they had simply “memorized” the answer to this widely circulated version of the puzzle. This became clear when the puzzle was modified with different dates: faced with these variations, the LLMs consistently failed to solve them, revealing their lack of proper logical understanding.
More tellingly, when asked to write programs to solve such puzzles, the LLMs couldn’t properly implement the logic of how each statement changes Albert and Bernard’s knowledge states. For example, one LLM wrote code that eliminated dates based on Albert’s first statement but failed to track how this elimination would affect Bernard’s subsequent reasoning. Another LLM recognized that Bernard’s statement meant he had figured out the date but couldn’t model how Bernard would have reached that conclusion based on Albert’s previous statement.
This limitation connects directly to what Chalmers identifies as necessary preconditions for consciousness: robust world models and self-models that can track changing states over time. The Cheryl’s Birthday puzzle reveals that current LLMs lack a crucial aspect of consciousness despite their impressive language capabilities e.g., the ability to maintain and update separate models of different minds’ knowledge states and reason about how these states influence each other over time. Just as the Voight-Kampff test reveals replicants’ struggles with emotional perspective-taking, this puzzle exposes LLMs’ inability to truly model and reason about other minds.
The evidence we’ve examined reveals a nuanced picture of machine consciousness. While LLMs demonstrate some aspects of conscious-like behavior, particularly in their ability to handle certain Theory of Mind tasks, they still fall short of human-like performance in crucial ways. The gap becomes especially apparent when we probe deeper with tests like Cheryl’s Birthday puzzle, which reveals their limitations in modeling and updating complex mental states. Chalmers’ framework helps us structure these observations, suggesting specific capabilities we look for in potentially conscious systems, from world models and self-models to more sophisticated forms of information processing.
Yet these investigations lead us back to fundamental questions about consciousness itself. As we observe increasingly sophisticated artificial systems, we must grapple with deeper philosophical puzzles: Does consciousness emerge gradually as systems become more complex, appearing at some threshold of computational sophistication? Or is consciousness, as some philosophers argue, an irreducible phenomenon that can’t be fully explained by computational processes alone? The answer might lie not just in understanding artificial minds but in better understanding what makes any biological or artificial system conscious in the first place.

Leave a Reply