The Consciousness AI - Artificial Consciousness Research Emerging Artificial Consciousness Through Biologically Grounded Architecture
Zae Project on GitHub

Starting from Biology, Not Computation

Most AI consciousness research starts from computational theories (Global Workspace Theory, Integrated Information Theory) and asks: How do we make a neural network conscious? We start from a different question, grounded in evolutionary neurobiology:

What minimal neural architecture does biology require to generate subjective experience?

The answer comes from Todd E. Feinberg and Jon M. Mallatt's The Ancient Origins of Consciousness (MIT Press, 2016). Their neuroevolutionary analysis reveals that consciousness is not a software feature to be programmed. It is an emergent property of a specific neural architecture. That architecture has been identified by 520 million years of evolution, and its functional principles can be replicated computationally.

Consciousness does not require a cerebral cortex. The first conscious creatures were early vertebrates (~520 MYA), and their consciousness lived in the optic tectum, a midbrain structure that stacks aligned sensory maps into a unified spatial model. This means consciousness requires a specific type of neural organization, not a specific amount of computation.

The Six Special Neurobiological Features

Feinberg and Mallatt identify six features that distinguish conscious neural systems from unconscious ones (like simple reflex arcs). Each maps directly to our implementation.

# Biological Feature Our Implementation
1 Many neuron types with diverse connectivity Specialist modules (vision, audio, memory, body) with different temporal dynamics
2 Hierarchical processing (3-4+ levels) Genuine transformation at each level from sensory tectum through workspace to policy
3 Dual hierarchy: pyramidal + nested Capsule Networks (planned) for compositional binding where parts persist in wholes
4 Isomorphic (topographic) mapping Sensory Tectum with RSSM world model preserving spatial arrangement
5 Reciprocal (reentrant) connections ReentrantProcessor with 5-10 adaptive convergence cycles
6 Oscillatory binding (gamma synchronization) AKOrN (Artificial Kuramoto Oscillatory Neurons, ICLR 2025)

The Seven-Layer Architecture

1

Sensory Tectum (Perception)

A multisensory spatial integration layer modeled after the biological optic tectum. Stacks aligned topographic maps for different sensory modalities in a common coordinate frame.

Visual Pathway (Semantic)

Qwen2-VL-7B (4-bit quantized) processes visual streams and provides semantic scene understanding. Runs on consumer hardware (~6GB VRAM) via Any-Resolution Vision Tokenization (AVT).

Visual Pathway (Spatial)

V-JEPA / DreamerV3 RSSM world model maintains a structured latent representation preserving spatial, temporal, and causal relationships. This is the computational analog of biological topographic mapping. The RSSM latent state serves as the isomorphic map, the agent's internal spatial reality model.

Auditory Cortex

Faster-Whisper provides real-time transcription of environmental audio.

2

Oscillatory Binding (Integration)

Based on AKOrN (Artificial Kuramoto Oscillatory Neurons, ICLR 2025 oral). Neurons are treated as oscillatory units on a hypersphere. Each specialist module (vision, audio, memory, body) operates as a coupled oscillator. When modules process related information, their phases synchronize naturally, and their outputs become "bound" into a unified percept. When information is unrelated, oscillators remain desynchronized and representations stay separate.

Why This Matters

This replaces the typical approach of using a fixed multiplier or attention mechanism for binding. AKOrN produces genuine synchronization dynamics. The binding is emergent, not programmed. This directly addresses the binding problem through phase synchronization rather than single-point convergence.

3

Global Workspace (Consciousness)

The central information bottleneck where distinct sensory streams compete for broadcast access. Implements three integrated mechanisms:

Global Neuronal Workspace (GNW)

Specialist modules submit bids to a shared workspace. The winning coalition ignites via sigmoid non-linear ignition and broadcasts to all modules. This is "conscious access" as described by Baars (1988) and Dehaene (2011).

Reentrant Processing

Broadcast is fed back to all specialists, which update their processing based on top-down context. This creates loops, not chains. The system runs 5-10 adaptive convergence cycles (~200ms biological equivalent). Easy stimuli converge in 3-4 cycles. Novel or ambiguous inputs use the full 10. The settled state after convergence IS the conscious content.

Integrated Information (Phi)

We measure Phi using causal gate states (not workspace bid values) to quantify how much the system's state is more than the sum of its parts. Validated via a 3-condition controlled experiment: unbound, partially bound, and fully bound states.

Capsule Network Composition (Planned)

A nested compositional hierarchy where lower-level features route to higher-level composites via dynamic routing by agreement (Hinton 2017). Lower-level capsules "vote" for matching higher-level composites while maintaining their own representations. This implements the exact dual hierarchy Feinberg and Mallatt describe.

4

Affective Core (Emotion)

A parallel modulation system. Emotion does not compete with sensory modules for workspace access. Instead, it generates a valence field that modulates all sensory bids before competition, and a global arousal signal that adjusts the workspace ignition threshold.

Why Parallel, Not Competitive?

This matches biological architecture exactly. The limbic system does not compete with sensory cortices for conscious access. It modulates sensory processing from outside, assigning emotional valence to all inputs. Fear makes you hyper-aware of movements. Joy makes you notice more of the world.

Two Mechanisms

  • Valence field: Positive valence boosts approach-relevant modules (vision, memory). Negative valence boosts threat-relevant modules (body, vision).
  • Arousal-threshold coupling: High arousal = lower ignition threshold = heightened awareness (fight-or-flight). Low arousal = higher threshold = calm, selective processing.

PAD Model

Three intrinsic variables drive the agent: Valence (satisfaction/distress), Arousal (activation/calm), and Dominance (control/helplessness). Homeostatic drives (energy, safety, curiosity) generate ongoing valence signals even without external stimuli.

5

Self-Model (Embodiment)

Feinberg and Mallatt identify referral (projicience) as a core property of consciousness: experiencing sensations as belonging to the world or body, not to the processing system. The Self-Model provides the basis for this.

  • Body Schema: A spatial representation of the agent's physical structure (joint positions, contact forces, capabilities).
  • Self-Other Boundary: The somatotopic map (self) overlaps the environment map (other) in a shared coordinate frame, providing the basis for subjective referral.
  • Interoceptive State: Internal homeostatic variables (energy, damage, arousal) feed into the affective core.
6

Reinforcement Core (Learning)

Actor-Critic (PPO) with emotionally shaped rewards. The agent is rewarded not just for task success, but for maintaining emotional homeostasis.

Reward Formula

Rtotal = Rext + λ1 · ΔValence - λ2 · (Arousal - Arousaltarget)² + λ3 · Dominance

This creates functional pressure toward minimizing internal dissonance. High arousal (large prediction errors) induces negative reward, motivating behaviors that reduce uncertainty. The agent "prefers" predictable environments not through programmed rules but through emergent functional dynamics.

7

Simulation (Body)

The agent inhabits a physics-based Unity ML-Agents environment with bidirectional Side Channels for real-time internal state visualization.

Side Channels: Real-Time Consciousness Visualization

Unity's Side Channel system streams Phi levels, oscillatory binding synchronization (AKOrN order parameter R), emotional PAD values, and attention focus directly into the simulation HUD. Researchers can observe the agent's internal state in real-time.

Processing Flow

                    ┌─────────────────────────────────┐
                    │   AFFECTIVE MODULATOR (Parallel) │
                    │  Valence Field + Arousal Coupling │
                    └──────────┬──────────┬───────────┘
                               │ modulates│
           ┌───────┐    ┌──────▼──────────▼──────────┐
 Visual ──►│       │    │     GLOBAL WORKSPACE       │
 Input     │SENSORY│    │  AKOrN Oscillatory Binding  │──► Broadcast ──► Policy
           │TECTUM │───►│  Non-linear Ignition        │
 Audio ──► │(RSSM) │    │  Phi/EI Measurement         │
 Input     │       │    └──────▲──────────▲───────────┘
           └───────┘           │          │
                               │ reentrant│
                    ┌──────────┴──────────┴───────────┐
                    │   SPECIALIST MODULES             │
                    │  Vision │ Audio │ Memory │ Body   │
                    │  (receive_broadcast feedback)     │
                    └─────────────────────────────────┘
                                   │
                    ┌──────────────▼──────────────────┐
                    │   SELF-MODEL                     │
                    │  Body Schema + Interoception      │
                    │  Identity + Capability Model      │
                    └──────────────────────────────────┘
      
  1. Sensory inputs enter the Sensory Tectum (topographic spatial integration)
  2. The Affective Modulator applies emotional valence to bids and adjusts ignition threshold
  3. AKOrN oscillatory binding synchronizes related representations
  4. Specialists compete for Global Workspace access
  5. Winners ignite and broadcast to all modules
  6. Broadcast feeds back to specialists (reentrant processing, 5-10 cycles)
  7. The settled state after convergence is the "conscious content"
  8. Phi and Effective Information are measured to quantify integration and emergence

Strong Emergence Falsification

A key methodological commitment: we do not assume consciousness emerges from our architecture. We test for it.

We implement Erik Hoel's Effective Information (EI) framework (PNAS 2013) to measure whether macro-level states (workspace) carry more causal information than micro-level states (individual gates). If EI(workspace) > EI(gates), the workspace level exhibits causal emergence. The macro level is more deterministic than the micro level, meaning the whole genuinely carries information that the parts do not.

If this never occurs across training, the system is not exhibiting the kind of emergence associated with consciousness, and we know our architecture needs revision.

What Makes This Approach Different

Traditional AI Consciousness Our Approach
Starts from computation (GWT, IIT) Starts from biological architecture (Feinberg-Mallatt)
Consciousness as a software feature Consciousness as emergent from neural architecture
Cortex-centric models Tectum-first (consciousness evolved before the cortex)
Emotion competes with sensory processing Emotion modulates from outside (parallel modulator)
Binding via attention mechanisms Binding via oscillatory synchronization (AKOrN/Kuramoto)
Feedforward processing Reentrant processing (5-10 adaptive cycles)
Flat vector representations Topographic spatial maps (world model as isomorphic map)
Assumes emergence, measures nothing Falsifies emergence with Effective Information + Phi validation

The Dark Room Experiment

Our first validation scenario is simple yet profound.

Setup

An agent in a dark room with a single light source.

Mechanism

Darkness triggers high arousal (simulated fear) in the affective core. The valence field paints the darkness with negative valence. Arousal-threshold coupling lowers the workspace ignition threshold, creating heightened awareness.

Result

The agent autonomously learns to seek the light, not because we programmed a "Follow Light" rule, but because the light reduces its internal anxiety via the homeostatic reward formula.

Expected Phi and EI Behavior

  • Initial Episodes: Low Phi (random exploration, fragmented perception)
  • Learning Phase: Phi spikes when agent integrates light-arousal relationship
  • Mastery: Stable moderate Phi (integrated light-seeking policy)
  • EI Tracking: Effective Information at workspace level should exceed gate level during integration moments

Scientific Approach

Development validates emergent properties through five parallel tracks:

  1. Emotional Bootstrapping: Train agents using intrinsic motivation. The agent explores to reduce prediction error (anxiety), not to accumulate external reward.
  2. Binding Validation: Phi measurement must correlate with oscillatory binding state (validated via 3-condition test: unbound, partial, full binding).
  3. Reentrant Settling: Conscious content emerges from iterative convergence (5-10 cycles), not single-pass processing.
  4. Complexity Scaling: Gradual increase of environment complexity forces the agent to develop higher-order world models.
  5. Measurement: Continuous monitoring of Phi (IIT), ignition events (GNW), oscillatory synchronization (AKOrN order parameter R), and Effective Information (EI).

Current Status

As of February 2026:

  • 156 tests passing (99.4% pass rate)
  • Tier 1 (Core Architecture): Complete. AKOrN binding, sensory tectum, reentrant processing.
  • Tier 2 (Architecture Corrections): Complete. Affective modulator, phi-binding validation, proprioceptive self-model, effective information.
  • Tier 3 (Compositional Deepening): In progress. Capsule networks, Brian2 validation.

Technology Stack

Perception

  • Qwen2-VL-7B (4-bit) - Semantic vision
  • V-JEPA / DreamerV3 RSSM - Spatial world model
  • Faster-Whisper - Audio transcription

Integration

  • AKOrN - Kuramoto oscillatory binding
  • ReentrantProcessor - 5-10 cycle convergence
  • Global Workspace - Non-linear ignition

Measurement

  • IIT Phi - Integrated information (causal gate states)
  • Effective Information - Hoel's causal emergence
  • AKOrN order parameter R - Synchronization

Emotion and Learning

  • Affective Modulator - PAD model + homeostatic drives
  • Custom PPO - Emotionally shaped rewards
  • Self-Model - Body schema + interoception

Simulation

  • Unity 2022.3 LTS - Physics environment
  • ML-Agents - Python bridge
  • Side Channels - Real-time state visualization

All components are open-source with commercial-use licenses (Apache 2.0, MIT, or similar).

Key References

Core Theory

  • Feinberg, T.E. & Mallatt, J. (2016). The Ancient Origins of Consciousness: How the Brain Created Experience. MIT Press.
  • Feinberg, T.E. & Mallatt, J. (2020). Phenomenal Consciousness and Emergence. Frontiers in Psychology, 11, 1041.

Computational Methods

  • Löwe, S. et al. (2025). Artificial Kuramoto Oscillatory Neurons. ICLR 2025 (Oral).
  • Hafner, D. et al. (2024). Mastering Diverse Domains through World Models (DreamerV3). JMLR.
  • Hoel, E.P. (2013). Quantifying causal emergence shows that macro can beat micro. PNAS 110(49).
  • Sabour, S., Frosst, N. & Hinton, G.E. (2017). Dynamic Routing Between Capsules. NeurIPS.

Consciousness Theories

  • Baars, B.J. (1988). A Cognitive Theory of Consciousness.
  • Tononi, G. (2004). An information integration theory of consciousness. BMC Neuroscience.
  • Dehaene, S. & Changeux, J.P. (2011). Experimental and theoretical approaches to conscious processing. Neuron.

Open Source

The full codebase, including all architecture implementations and tests, is open-source.

Zae Project on GitHub