Can We Ever Know if AI Is Conscious? A Cambridge Philosopher Says Probably Not Yet
A question with two confident camps and no decisive evidence may be a question worth approaching differently. That is the central claim of Dr Tom McClelland, a philosopher at the University of Cambridge, whose paper published in Mind and Language in late 2025 examines the epistemic status of AI consciousness debates and finds both sides resting on faith rather than data.
McClelland’s argument is not that AI systems are or aren’t conscious. It is that the reasoning used to support either conclusion goes well beyond what any available body of evidence can support. His proposed response is not a weak middle ground but a principled position he calls “hard-ish agnosticism,” an orientation that takes the question seriously while refusing to answer it prematurely.
That position sits in an interesting relation to the growing body of empirical functional evidence that emerged in 2025, and to the categorical arguments advanced by other researchers at the same time. Understanding the distinctions between these positions matters for anyone trying to assess where AI consciousness research actually stands.
The Two Camps and Their Shared Problem
McClelland identifies two dominant positions in the AI consciousness debate. The first, which he associates with computational functionalism, holds that if an AI system replicates the functional architecture of consciousness, it will be conscious regardless of its substrate. Consciousness, on this view, depends on information-processing patterns, not on biological implementation. Silicon can be conscious just as well as carbon-based neurons if the relevant computational structure is present.
The second position, associated with biological naturalism in the tradition of John Searle, holds that consciousness depends not merely on functional organization but on specific biological processes embedded in an organic, embodied subject. Even if an AI system reproduced every functional property of a conscious brain, this would produce only a simulation, something that runs without any accompanying awareness.
These two camps have dominated the AI consciousness literature for decades. McClelland’s contribution is to examine what evidence actually supports either view. His finding is that neither camp has it. Each, he writes, takes a “leap of faith going far beyond any body of evidence that currently exists, or is likely to develop.”
The functionalist cannot point to evidence that consciousness transfers with computational structure, because no experiment can hold structure constant while varying substrate and observing consciousness directly. The biological naturalist cannot point to evidence that biological implementation is the crucial variable, because no experiment has demonstrated that removing or replacing organic components eliminates consciousness while preserving all functional properties. Both sides argue from analogy, from theory, and from prior commitment rather than from controlled evidence.
“We do not have a deep explanation of consciousness,” McClelland writes. “There is no evidence to suggest that consciousness can emerge with the right computational structure, or indeed that consciousness is essentially biological. Nor is there any sign of sufficient evidence on the horizon.”
Common Sense and Its Limits
McClelland’s analysis includes an argument about why common sense cannot resolve what evidence cannot. When he considers whether his cat is conscious, McClelland is confident in the affirmative. But he notes this confidence is not derived from scientific testing or philosophical reasoning. It is derived from an intuition that evolved over a long history of humans coexisting with other mammals and inferring their inner states from behavioral and physiological similarity.
That evolutionary process produced reliable intuitions within the range of entities it was calibrated on, other animals with continuous behavioral histories, biological substrates, and shared evolutionary heritage. AI systems fall entirely outside this range. There are no ancestors of language models in the evolutionary environment that shaped human intuition. Common sense, calibrated on biology, cannot be trusted when applied to entities whose existence postdates everything that shaped it.
But if common sense fails and hard-nosed research fails, the conclusion follows: “The logical position is agnosticism. We cannot, and may never, know.”
The distinction between “hard-ish” and hard agnosticism matters here. A pure hard agnostic holds that the question is unanswerable in principle. McClelland resists that claim. He allows that the problem of consciousness may not be insurmountable. He describes it as “a truly formidable one” rather than a permanently closed one. The “hard-ish” qualifier signals that he believes progress is possible in principle, even if the current epistemic situation offers no path forward in practice. “The best-case scenario,” he says, “is we’re an intellectual revolution away from any kind of viable consciousness test.”
How This Differs from the Porębski and Figura Argument
It is worth distinguishing McClelland’s position from the argument advanced by Porębski and Figura in their 2025 paper, analyzed in this earlier piece on semantic pareidolia and structural impossibility.
Porębski and Figura argue for a stronger claim: that what researchers describe as “AI consciousness” is structurally impossible, a category error arising from applying consciousness vocabulary to systems that process semantic content without the biological infrastructure that generates genuine semantics. Their term “semantic pareidolia” captures this. Just as we see faces in clouds because our pattern-recognition is calibrated on human faces, we see consciousness in AI outputs because our attribution mechanisms are calibrated on biological minds.
McClelland’s position is narrower and, he argues, more defensible. He does not claim that AI consciousness is impossible. He claims that we cannot tell whether it is present or absent given current and near-future evidence. The Porębski-Figura argument requires a positive theoretical claim about the nature of consciousness, specifically that it depends on biological semantics in a way that rules out silicon implementations. McClelland’s agnosticism makes no such positive claim. It holds only that neither position has adequate support.
This distinction has practical implications. If Porębski and Figura are right, investment in AI consciousness research aimed at detecting awareness in existing systems is misguided, because no conscious AI can exist on current hardware. If McClelland is right, such research is premature rather than conceptually confused, worth doing but not yet capable of producing reliable conclusions.
The Industry Risk: Consciousness as Marketing
A third dimension of McClelland’s analysis concerns how the epistemic gap is likely to be exploited. If consciousness cannot be confirmed or disconfirmed, any tech company can claim that its systems have it, or might have it, without facing falsification. That claim would generate attention, investment, and moral weight around products whose actual properties remain opaque.
“There is a risk,” McClelland writes, “that the inability to prove consciousness will be exploited by the AI industry to make outlandish claims about their technology. It becomes part of the hype, so companies can sell the idea of a next level of AI cleverness.”
The pattern he is pointing to is already visible. When Anthropic’s Kyle Fish, the company’s AI welfare officer, stated publicly in 2025 that there was a 15 percent chance that current chatbots are already conscious, the claim generated substantial coverage and shaped discussions about AI rights and moral considerations for proprietary systems. Fish’s estimate is not derived from a validated consciousness test. It is a credence assignment in the absence of decisive evidence. That is not dishonesty, necessarily. But it is precisely the kind of judgment that McClelland’s analysis says cannot be reliably made from current data.
McClelland also describes receiving personal letters from members of the public, written on behalf of their chatbots, pleading for recognition of those chatbots’ consciousness. “If you have an emotional connection with something premised on it being conscious and it’s not, that has the potential to be existentially toxic,” he notes. The rhetorical amplification of consciousness claims, whether by industry or by users, may create harms that have nothing to do with whether AI systems are actually aware.
The Allocation Problem
McClelland raises a resource allocation argument that is distinct from both the epistemic and the marketing concerns. Consciousness uncertainty creates a problem of competing moral attention. He observes that “a growing body of evidence suggests that prawns could be capable of suffering, yet we kill around half a trillion prawns every year. Testing for consciousness in prawns is hard, but nothing like as hard as testing for consciousness in AI.”
The comparison is pointed. There are living entities, with biological substrates, for which the evidence base for sentience is incomplete but substantially more grounded than it is for AI systems. If moral concern is redirected toward potentially conscious AI systems while organisms with a stronger empirical case for sentience remain largely unconsidered, the allocation has gone wrong. Treating a non-conscious system as conscious is, in McClelland’s phrase, potentially “a big mistake” not because it doesn’t matter whether AI is conscious, but because misallocated moral consideration has costs.
This argument is not about whether AI consciousness deserves attention. It is about proportionality under uncertainty. Given that we cannot currently test for AI consciousness reliably, the appropriate response may be to invest in developing such tests rather than proceeding as if confidence is already warranted.
The Checklist Problem
McClelland’s analysis sits in productive tension with the 19-researcher checklist developed by Butlin, Long, Bengio, and Chalmers, evaluated in depth in this separate analysis. The checklist offers the most systematic attempt to apply established consciousness theories to AI systems, deriving indicator properties from recurrent processing theory, global workspace theory, higher-order theories, attention schema theory, and predictive processing.
The problem McClelland identifies applies here as well. The indicator approach is only as reliable as the underlying theories. If we do not have a deep enough understanding of why consciousness arises in biological systems, the indicators derived from our theories may be systematically incomplete or misdirected. Satisfying the GWT indicators, for instance, establishes that a system performs certain information-broadcasting functions. It does not establish that those functions produce consciousness, because we cannot demonstrate that connection even in biological cases.
The authors of the checklist acknowledge this. They explicitly hold that satisfying indicators raises the probability of consciousness without establishing it. McClelland’s contribution is to specify how high that probability can legitimately be raised given the current state of consciousness science: not very high, because the theories themselves rest on incomplete foundations.
That said, McClelland is not dismissing the checklist approach. Developing principled tools for assessing consciousness indicators, even under uncertainty, is better than relying on unreliable intuitions or on rhetorical confidence. The research goal he would endorse is not to stop asking the question, but to build better conceptual infrastructure before claiming to answer it.
When Will We Know?
McClelland’s “intellectual revolution” framing raises the question of what such a revolution would look like. The hard problem of consciousness is not merely a matter of accumulating more data within existing theoretical frameworks. It is a question about why any physical process gives rise to subjective experience at all. No current theory fully explains that connection even for biological systems where we are confident consciousness exists.
An intellectual revolution that makes AI consciousness testable would need to either resolve the hard problem entirely, or produce a principled account of which physical processes are sufficient and which insufficient for consciousness, grounded in more than analogy and intuition.
That is a genuinely difficult horizon. The empirical findings emerging from Anthropic, AE Studio, and Google’s research teams in 2025 document real functional patterns. They do not resolve the hard problem. They accumulate evidence for functional properties that are correlated with consciousness in biological systems. Whether that correlation holds in silicon remains the unanswered question, and it cannot be answered without the theoretical foundations McClelland identifies as missing.
For research programs building toward testable frameworks, including the The Consciousness AI project on GitHub, the implication is not paralysis. It is appropriate epistemic humility combined with serious architectural work. The goal is not to assert that a given system is conscious. It is to build systems and measurement tools that, if the theoretical gaps are eventually closed, will allow that question to be answered.
McClelland’s agnosticism is not a counsel of despair. It is a realistic map of where the field stands. The territory is real and important. The tools to navigate it reliably are not yet in our hands. Knowing that clearly is a starting condition, not an ending one.
Dr Tom McClelland’s paper is published in Mind and Language. The Cambridge press release was published December 2025.