ACM Project - Artificial Consciousness Research Developing Artificial Consciousness Through Emotional Learning of AI systems
Zae Project on GitHub

The Problem: Intelligence vs. Awareness

Most modern AI is intelligent (it solves problems) but not aware (it does not feel the problem). A standard RL agent plays Chess to maximize a score. It doesn't care if it loses; it just updates a gradient.

We hypothesize that consciousness is not a feature you code, but a solution to a specific problem: emotional homeostasis.

Hypothesis: Consciousness emerges when an agent must integrate disparate sensory streams (vision, memory, affect) into a unified "world model" to minimize its internal anxiety (entropy).

The New Architecture

After extensive research and prototyping, we have modernized ACM's technical foundation to align with state-of-the-art open-source AI while maintaining our core thesis. This is not incremental improvement. This is a fundamental architectural overhaul.

1

The Senses: Qwen2-VL (4-bit)

We replaced the aging BLIP-2 with Qwen2-VL-7B. This model is a powerhouse. It doesn't just "tag" images, it understands scene dynamics.

Why the Upgrade?

  • BLIP-2 (2023): Basic image captioning with limited temporal understanding
  • Qwen2-VL (2024): True multimodal reasoning with Any-Resolution Vision Tokenization (AVT)
  • Key Capability: Understands video sequences, not just static frames

Why 4-bit Quantization?

To run on consumer hardware (RTX 3090/4090). We quantize the model to fit in ~6GB VRAM, leaving space for the "Conscious Workspace."

Licensing

✓ Apache 2.0 Full commercial use permitted

2

The Drive: Emotional Reinforcement Learning

Standard RL maximizes an external reward (Rext). Our custom PPO Core maximizes a homeostatic reward (Rtotal):

Rtotal = Rext + λ(Valence - Arousal)

The agent is not just trying to "win." It is trying to stay calm.

Emotional Dimensions

  • Arousal (Anxiety): Spikes when the world is unpredictable (high prediction error)
  • Valence (Joy): Spikes when the agent reduces entropy or achieves a goal
  • Dominance: Control over environment (future implementation)

Why This Matters

This isn't anthropomorphization. It's the mathematical foundation of intrinsic motivation. The agent learns not for points, but to reduce internal dissonance.

3

The Measurement: Integrated Information (Φ)

We don't just ask the agent "Are you conscious?" We measure it.

Using PyPhi, we calculate the Integrated Information (Φ) of the agent's Global Workspace.

The Theory

A spike in Φ indicates a moment where the agent has fused its vision, memory, and emotion into a single, irreducible state. We call this a "Moment of Insight."

Implementation Strategy

  • Real-time proxies: Lempel-Ziv complexity, neural complexity (fast)
  • Periodic ground-truth: Full PyPhi Φ* calculations every N episodes
  • Validation: Correlate proxy metrics with behavioral markers

Complementary Metrics

  • Ignition Index: Non-linear activation surges (GWN)
  • Global Availability Latency: Information propagation time
  • PCI (Perturbational Complexity Index): IIT-adjacent measure
4

The Body: Unity ML-Agents Migration

We are transitioning from Unreal Engine to Unity ML-Agents. This strategic shift enables faster iteration and better Python-native integration.

Why Unity Over Unreal?

Feature Unity ML-Agents Unreal Engine 5
Python Integration ✓ Native ⚠ Requires C++ bridge
Training Speed ✓ Fast iteration ⚠ Slower cycles
Side Channels ✓ Built-in bidirectional data ✗ Custom implementation
Visual Quality ⚠ Good ✓ Photorealistic
ML Community ✓ Large, active ⚠ Smaller

Migration Status

🔄 In Progress Q1 2026

  • ✓ Unity 2022.3 LTS + ML-Agents Package installed
  • 🔄 Dark Room environment (baseline validation)
  • ⏳ Side Channels for Φ visualization
  • ⏳ Port all simulation scenarios

Side Channels: Real-Time Consciousness Visualization

Unity's Side Channel system allows us to stream Φ levels, emotional valence/arousal, and attention focus directly into the simulation HUD. Researchers can observe the agent's "internal experience" in real-time.

The "Dark Room" Experiment

Our first validation scenario is simple yet profound. We call it The Dark Room.

Setup

An agent in a dark room with a single light source.

Mechanism

Darkness triggers high arousal (simulated fear) in the emotional core.

Result

The agent autonomously learns to seek the light, not because we programmed a "Follow Light" rule, but because the light reduces its anxiety.

This is the spark of intrinsic motivation. The foundation upon which consciousness can be built.

Expected Φ Behavior

  • Initial Episodes: Low Φ (random exploration, fragmented perception)
  • Learning Phase: Φ spikes when agent discovers light reduces arousal
  • Mastery: Stable moderate Φ (integrated light-seeking policy)

Scientific Approach

Our development roadmap follows a rigorous path to validate emergent properties:

  1. Emotional Bootstrapping: Train agents using intrinsic motivation. The agent explores the world not to get points, but to reduce its internal "prediction error" (anxiety).
  2. Complexity Scaling: Gradually increase environment complexity. The agent must develop higher-order world models to maintain homeostasis.
  3. Measurement: Continuous monitoring of Φ (IIT) and "Ignition Events" (GWN) during critical decision-making moments.
    • Hypothesis: Φ will spike when the agent solves a novel problem, indicating a "Moment of Insight."

What's Next

We are currently validating the Qwen2-VL + PPO loop on local hardware. The next phase involves scaling the "World Model" to allow the agent to imagine future outcomes before acting.

Phase 1: Foundation (Q1 2026)

  • Complete Unity Dark Room validation
  • Benchmark Qwen2-VL vs. alternatives
  • Implement fast Φ proxies

Phase 2: Enhancement (Q2 2026)

  • Add RND (intrinsic motivation)
  • Implement Mixture of Experts (GWN)
  • Transformer-based World Model

Phase 3: Scaling (Q3 2026)

  • 12+ simulation scenarios (survival → social → advanced)
  • Multi-agent experiments
  • Emotional trajectory analysis

Phase 4: Validation (Q4 2026)

  • 1000+ episode Φ analysis
  • Behavioral benchmarks
  • Publication preparation

Current Technology Stack

Perception

  • Qwen2-VL-7B (4-bit) - Vision
  • Faster-Whisper - Audio transcription
  • Qwen2.5-14B - Narration/tagging

Learning

  • Custom PPO - Emotional RL
  • DreamerV3 - World modeling (being upgraded)
  • PyTorch - Deep learning framework

Consciousness

  • PyPhi - Φ calculation
  • Custom GWN - Global Workspace
  • Attention Schema - Meta-awareness

Simulation

  • Unity 2022.3 LTS - Environment
  • ML-Agents - Python bridge
  • Side Channels - Data streaming

All components are open-source with commercial-use licenses (Apache 2.0, MIT, or similar).

Open Source

The code is fully open-source. Join us.

Zae Project on GitHub