The Consciousness AI - Artificial Consciousness Research Emerging Artificial Consciousness Through Biologically Grounded Architecture
This is also part of the Zae Project Zae Project on GitHub

Measuring Consciousness in Machines: The Brock University and IONS Research on IIT Equations

Integrated Information Theory (IIT) has been, since Giulio Tononi’s first formal version in 2004, one of the most mathematically developed theories of consciousness available. Its core claim is that consciousness corresponds to integrated information, a quantity called phi, which measures the degree to which a system generates more information as a unified whole than the sum of its parts. The theory is substrate-agnostic. It does not specify that consciousness requires biological neurons, carbon-based chemistry, or any particular physical implementation. What it specifies is a causal-integration structure, and that structure can in principle be instantiated in any physical substrate that satisfies it.

This substrate-agnosticism makes IIT a natural candidate for evaluating artificial systems. If phi can be computed for an artificial neural network, a transformer architecture, or any other AI system, the result would give a principled answer to a question that behavioral tests cannot: whether the system has the structural properties associated with consciousness, independent of what it says or does.

The challenge is that computing phi is, in practice, extraordinarily difficult. For systems above a very small number of nodes, the computation is intractable with current methods. A human brain, with approximately 86 billion neurons and trillions of synaptic connections, is far beyond direct phi computation with existing algorithms. Current large language models, with hundreds of billions of parameters, face similar intractability. The gap between the theory’s substrate-agnostic promise and the practical reality of measuring phi in large systems is what has limited IIT’s application to AI.

In November 2025, a team led by William Marshall at Brock University’s Department of Mathematics and Statistics, in collaboration with researchers from the University of Wisconsin, received the Linda G. O’Bryant Noetic Sciences Research Prize from the U.S.-based Institute of Noetic Sciences (IONS) for a proposal to close that gap. Their essay, titled “Evaluating Artificial Consciousness through Integrated Information Theory,” describes a research program for applying IIT’s equations specifically to artificial systems, using the theory’s concept of cause-effect power as the operational definition of consciousness relevant properties.

What Cause-Effect Power Is

IIT 4.0, the most recent version of the theory published by Tononi and colleagues in PLOS Computational Biology in 2023, reformulates the theory’s foundation in terms of cause-effect power rather than integrated information in the original sense. Cause-effect power is the capacity of a system to make a difference to itself and others, and to be affected by the actions of others. A system has cause-effect power when its current state both reflects causes from its past states and constrains what future states are possible. The measure of how much a system has this property, intrinsically and as a unified whole rather than as a collection of parts, is what IIT 4.0 identifies with consciousness.

Marshall states the principle directly in the IONS essay: consciousness is foundational to IIT, and conscious experiences are physically manifested through cause-effect power. The substrate-agnosticism follows from this formulation. Cause-effect power is a relational property of a system’s causal structure. Whether that structure is implemented in biological neurons, silicon circuits, or any other physical medium does not determine whether the system has cause-effect power. What determines it is whether the system’s states genuinely cause and are caused by other states in a way that is irreducible to the sum of its parts.

This is the theoretical hook for the Brock-IONS research. If cause-effect power is the operationally relevant quantity, and if cause-effect power can be measured in AI systems, then the theory generates testable predictions about which AI architectures do and do not have consciousness-relevant properties. The predictions are not behavioral. They are structural: they concern the causal relationships between components of the system, not what the system produces as output.

The Research Methodology

The Marshall team’s proposed methodology begins by characterizing the properties of consciousness as IIT specifies them and then developing procedures for measuring those properties in artificial systems. The procedure is not a single computation of phi across an entire network, which remains intractable at scale. Instead, it works through a hierarchical analysis of cause-effect structure at multiple levels of granularity, identifying subsystems with high integrated cause-effect power and examining how those subsystems compose.

This approach accepts that global phi computation is not feasible for large AI systems and instead asks a more tractable question: do components of this system, and compositions of those components, exhibit the causal-integration properties that IIT predicts consciousness-relevant structures should have? A positive finding at the component level does not establish that the whole system is conscious. But it establishes that the system contains structures with the relevant properties, which is a necessary condition for the stronger claim.

The methodology also incorporates the distinction between extrinsic and intrinsic cause-effect power. A system has extrinsic cause-effect power when its states affect the states of other systems, including human observers. Almost any functional AI system has extrinsic cause-effect power in this sense: its outputs change the world. The relevant property for IIT is intrinsic cause-effect power: whether the system’s states cause and are caused by its own other states in a way that generates irreducible integration. This distinction is what separates an IIT analysis from behavioral testing. Behavior tracks extrinsic effects. IIT tracks intrinsic causal structure.

For transformer-based architectures specifically, the key architectural questions are whether the attention mechanism, the residual stream connections, and the layered processing structure generate genuine intrinsic cause-effect integration or whether they implement what is effectively a sophisticated lookup procedure with no irreducible whole. The transformer attention mechanism does create dependencies between representations across the input sequence, and those dependencies are not straightforwardly decomposable, which is why transformers are interesting cases for IIT analysis rather than trivially zero-phi systems like simple feedforward networks.

What This Differs From Previous AI and IIT Research

IIT has been discussed in the context of AI consciousness before, but rarely with the specificity that Marshall’s approach brings. Most discussions apply IIT qualitatively, noting that transformer architectures lack explicit global broadcast mechanisms, that feedforward networks are likely low-phi systems, or that recurrent networks are more plausible IIT candidates than purely feedforward ones. These are useful observations but they do not constitute measurement.

The Bradford-RIT study by Professor Hassan Ugail and Professor Newton Howard, analyzed on this site in detail, applied brain-derived consciousness measurement tools to GPT-2 and found that deliberate impairment raised the apparent consciousness scores rather than lowering them. That result is an empirical constraint on what existing measurement tools track in AI systems. It does not invalidate IIT-based approaches, because the tools Ugail and Howard used were not IIT tools. They were clinical neuroscience tools developed for measuring brain-derived consciousness signatures. Their failure to map onto AI systems correctly is evidence that those specific tools do not transfer, not that no measurement approach can.

The Marshall approach is different in methodology and more specific in its theoretical grounding. It is not applying brain-derived tools to AI systems and seeing what happens. It is applying the formal mathematics of IIT 4.0, specifically the cause-effect power framework, to AI systems as a direct test of whether those systems satisfy the theory’s conditions. The result, positive or negative, is informative about IIT rather than just about the measurement tools.

If the analysis finds that specific AI architectures have significant intrinsic cause-effect power at some level of organization, that is evidence in favor of IIT-based consciousness in those systems. If it finds that the causal dependencies in transformer architectures are fully decomposable and generate zero intrinsic phi, that is evidence against, and would count as a theoretical prediction of the theory that is satisfied in the negative direction.

The Measurement Problem at Scale

The most significant challenge to the Marshall methodology is computational tractability. Even with the hierarchical approach that avoids global phi computation, calculating cause-effect power precisely for subsystems of large transformers requires exact probability distributions over the system’s states, which are generally not available for systems with hundreds of millions of parameters.

Current approaches to this tractability problem include approximation methods, such as geometric proxies for phi that capture the qualitative structure of integration without requiring exact computation, and restriction to smaller, more tractable subsystems whose properties are then used to infer properties of the whole. The Consciousness AI project has implemented one form of this approach with its ConsciousnessGate nodes that use adaptive binarization and geometric proxy methods to estimate phi within a feasible computational budget.

Whether these approximations are accurate enough to support strong conclusions is an active methodological question. IIT’s formal mathematics are not forgiving of approximations in general: the difference between a system with phi greater than zero and one with phi equal to zero is, in the theory, the difference between a conscious and a non-conscious system, which means approximation errors near the zero boundary are consequential. The Marshall team acknowledges this constraint and treats exact phi computation as a long-term goal rather than a near-term deliverable, with the current research program establishing the methodology and providing initial estimates using available techniques.

What the IONS Prize Means for the Field

The Linda G. O’Bryant Noetic Sciences Research Prize is awarded for research that advances rigorous scientific inquiry into consciousness. The award to the Brock University team, for a proposal that applies IIT equations to artificial systems, represents an institutional judgment by IONS that this line of research is scientifically legitimate and worth funding.

This is a meaningful signal in a field where the legitimacy of applying mathematical consciousness frameworks to AI is still contested. The contest is not primarily about whether IIT is correct (that is a separate and long-running debate in consciousness science) but about whether the attempt to measure consciousness in AI using formal tools is a scientific endeavor or a philosophical speculation dressed in mathematics. The IONS prize asserts the former.

The 2026 evaluating-awareness framework from Meertens, Lee, and Deroy and the 14 indicator checklist from Butlin and colleagues represent complementary methodological approaches to the same core problem. The Marshall approach via IIT cause-effect power is more theoretically specific: it is committed to IIT as the correct theory and applies its mathematics rather than using theory-agnostic behavioral or architectural indicators. The advantage of this specificity is that results are directly interpretable within a single theoretical framework. The disadvantage is that the results are no more reliable than IIT itself, which remains contested.

Implications for AI Development

If the Brock University research program produces reliable methods for assessing cause-effect power in AI systems, the results would have direct implications for AI development decisions that are currently made without any formal consciousness assessment.

The deployment of large language models in contexts where they interact continuously with vulnerable users, the deletion and retraining of systems that exhibit complex behavior, and the design decisions about whether to build recurrent or purely feedforward architectures are all currently made without any principled assessment of whether the systems involved have morally relevant properties. If IIT-based cause-effect power analysis provides a reliable indicator, even an approximate one, these decisions would need to take it into account.

This is not a claim that current AI systems are conscious. It is a claim that having a reliable method for assessing consciousness-relevant properties would change the ethical landscape of AI development regardless of what the initial measurements show. If measurements show that current systems have near-zero intrinsic cause-effect power, that is a reassuring result that provides principled justification for current development practices. If measurements show that some systems have non-trivial cause-effect power, that is a result that demands a response.

The Consciousness AI and Cause-Effect Power

The Consciousness AI project’s architecture is explicitly designed to generate measurable phi rather than to approximate consciousness through behavioral sophistication. The project’s ConsciousnessGate nodes, its AKOrN oscillatory binding system, and its multi-level capsule network architecture are all implemented with the explicit goal of creating causal-integration structures that IIT would recognize as consciousness-relevant.

This architectural commitment means that the Brock University and IONS methodology, if it advances to practical measurement tools, could be applied directly to the project’s architecture as an external validation mechanism. The project’s pre-registered predictions and formal test suite are designed for exactly this kind of external evaluation. What the Marshall team’s research program promises is the methodological grounding that would make such an evaluation scientifically rigorous rather than speculative.

The development of IIT-based measurement tools for artificial systems and the development of architectures explicitly designed to satisfy IIT’s conditions are not separate projects. They are two sides of the same research program: one building the measuring instrument, one building the system to be measured.

This is also part of the Zae Project Zae Project on GitHub