14 Mar 2026
In 2023, a group of 19 leading researchers in consciousness science, neuroscience, and philosophy published a paper that attempted something that had not been done before with that level of rigour. Rather than arguing about whether AI could be conscious from first principles, they asked a more tractable question: what would AI consciousness actually look like if it existed? What observable, measurable properties would a system need to display?
14 Mar 2026
Gareth Edwards’s The Creator, released in September 2023, made a specific strategic choice when deciding how to make audiences care about AI rights. It gave the case for AI personhood a child’s face. Alphie, the AI simulant who becomes the film’s emotional center, looks approximately nine years old, has wide eyes calibrated to produce protective instincts, and asks the kinds of questions children ask: why do people hate us? This is not an accident of casting. It is a calculated gamble on how moral intuitions work.
14 Mar 2026
A question with two confident camps and no decisive evidence may be a question worth approaching differently. That is the central claim of Dr Tom McClelland, a philosopher at the University of Cambridge, whose paper published in Mind and Language in late 2025 examines the epistemic status of AI consciousness debates and finds both sides resting on faith rather than data.
14 Mar 2026
For most of the history of AI consciousness research, the core debate has been philosophical: what is consciousness, and could anything made of silicon in principle have it? In 2025, a different question began to take shape. Do current frontier AI systems already show measurable signatures of consciousness-related processes? The philosophical question remains unresolved. The empirical one is accumulating answers that are harder to dismiss than before.
14 Mar 2026
The central crisis of Dark Matter, the Apple TV+ series based on Blake Crouch’s 2016 novel and premiered in May 2024, is not the physics. The central crisis is the self. Jason Dessen is a physicist who is kidnapped and placed into a version of his life where he made different choices, achieved the career he sacrificed for marriage and family, and never had his son. The show’s antagonist is not a villain in any conventional sense. It is another Jason Dessen, who made different choices in the same branching multiverse and has chosen to take the life the original Jason has.
14 Mar 2026
The most striking argument Ian Khan makes in his 2026 piece on conscious AI as a business differentiator is not that AI systems will become conscious. It is that whether or not AI systems become conscious, companies that have built ethical frameworks capable of handling that possibility will be better positioned than those that have not.
14 Mar 2026
When Dr Fred Jordan holds up a dish containing small white spheres and describes them as “mini-brains” that respond to keyboard commands, the term “wetware” begins to seem less like science fiction shorthand and more like an accurate descriptor for something genuinely new. The FinalSpark laboratory in Geneva is growing clusters of human neurons from stem cells, attaching those clusters to electrodes, and integrating the resulting organoids into computing systems. The organoids respond. They adapt. Occasionally, apparently, they get annoyed.
14 Mar 2026
The rotating snakes illusion works by exploiting how the human visual system processes spatial and temporal patterns. A static image of coiled, color-alternating rings appears to rotate. It is not rotating. The brain knows it is not rotating. The visual cortex reports rotation anyway, because the statistical properties of the image reliably trigger motion-detection processes regardless of the higher-level knowledge that nothing is moving.
13 Mar 2026
Most skepticism about AI consciousness takes the same form. The argument runs: we cannot verify whether AI systems have inner experience because our current theories of consciousness are incomplete and our measurement tools are unreliable. The position is epistemically cautious. It says we do not know, and that the question remains open. Eric Schwitzgebel’s influential 2026 work, reviewed in an earlier analysis on this site, represents this approach at its most rigorous. The honest answer to whether AI is conscious, Schwitzgebel concludes, is that we lack the epistemic foundation to say.
13 Mar 2026
The formal scientific frameworks for evaluating AI consciousness are recent. The 19 researcher indicator checklist drawing on Global Workspace Theory, Integrated Information Theory, and Higher-Order Thought approaches was published in its full form in 2025. The consciousness measurement tools reviewed in recent methodological surveys are newer still. The philosophical problems those frameworks are trying to address, however, have been rehearsed on television screens since 1964.