← All Experiments

AI vs GAN

Planned
AdversarialMeasurementValidation

Hypothesis

A generative adversarial network trained to mimic the computational signatures of consciousness (high Phi, global workspace dynamics, higher-order representations) will fail to reproduce certain structural properties of genuinely integrated systems — revealing which markers are reliable indicators of consciousness-like processing and which are superficially fakeable.

Overview

Before we can study consciousness in AI systems, we need measurement tools we can trust. This experiment is about validating those tools through adversarial testing.

The logic borrows from an old challenge in consciousness science: the "zombie problem." Could a system behave exactly like a conscious being without being conscious? In philosophy, this remains an open question. In engineering, we can make it empirical.

By training a GAN to fake consciousness markers, we discover which markers are robust — which ones carry information about the underlying causal structure of a system rather than just its surface behavior. Markers that resist adversarial imitation are the ones we should trust; markers that are easily faked are the ones we should be skeptical of.

This experiment serves a critical methodological function for the entire research program. Every other experiment relies on measuring consciousness-related properties in AI systems. If those measurements can be trivially gamed, our conclusions are worthless. AI vs GAN is how we pressure-test our instruments.

Methodology

  1. Catalog the computational signatures proposed as consciousness markers: Phi values, global workspace activation patterns, higher-order representation structures, attention dynamics, and information integration patterns.
  2. Build a generator network trained to produce activations and outputs that mimic these signatures without possessing the underlying causal structure that produces them naturally.
  3. Build a discriminator network trained to distinguish genuine consciousness markers from generated fakes.
  4. Analyze which markers the GAN successfully fakes and which resist imitation — these imitation-resistant markers are candidates for reliable consciousness indicators.
  5. Test whether the difficulty of faking a marker correlates with its theoretical importance in consciousness theories (e.g., IIT predicts that Phi cannot be faked by a feedforward system).
  6. Develop a robustness taxonomy of consciousness markers ranked by their resistance to adversarial imitation.

Status

Planned

This experiment is in the design phase. We are currently developing the detailed experimental protocol, identifying collaborators, and securing compute resources.


← Back to all experiments