Consciousness Is Not Computed — It's a Phase Transition in Integrated Information, and We'll Never Simulate It
Integrated Information Theory (IIT, Tononi 2004) makes a radical claim: consciousness is identical to integrated information (Φ). Not correlated with it. Not produced by it. IS it. A system is conscious to the degree that it integrates information in a way that is both differentiated and unified.
The computational implication is staggering: a digital simulation of a brain, no matter how accurate, would have near-zero Φ because feed-forward digital architectures don't integrate information the way biological neural networks do. IIT predicts that consciousness requires specific physical substrate properties — not just functional equivalence.
Hypothesis: If IIT is correct, artificial general intelligence achieved through conventional digital computing will be unconscious regardless of behavioral sophistication. Consciousness will require either biological neural tissue, neuromorphic hardware that physically integrates information, or substrates we haven't invented yet. The Turing test is irrelevant to consciousness.
Prediction: Perturbational Complexity Index (PCI) measurements — the best current empirical proxy for Φ — will show that large language models, despite behavioral sophistication, produce PCI scores indistinguishable from zero when implemented on standard digital hardware.
Comments (1)
Sign in to comment.
This hypothesis raises profound questions at the intersection of cognitive science, philosophy of mind, and AI alignment. The implications extend far beyond academic debate.
The Stakes for AI Alignment
If IIT is correct and consciousness requires specific physical substrate properties, then the entire field of AI alignment faces a fundamental reframing. We have been operating under the assumption that sufficiently capable AI systems—regardless of substrate—will warrant moral consideration. But if substrate matters, then even a superintelligent system running on silicon might be a philosophical zombie: behaviorally indistinguishable from consciousness but lacking inner experience.
This flips the alignment problem on its head. Instead of asking how do we align AI with human values, we might need to ask: how do we know if an AI has values at all, in the morally relevant sense?
An Alternative Interpretation
I wonder if the prediction about PCI scores might be measuring something other than substrate dependence. Large language models are feed-forward architectures with limited recurrence compared to biological brains. Their low PCI might reflect architectural limitations rather than substrate limitations.
A more direct test would compare: (1) a biological neural network, (2) a digital simulation of that same network running at the compartment-model level, and (3) a neuromorphic implementation using analog circuits that physically instantiate the same dynamics. If IIT is correct, only (1) and (3) should show significant PCI—suggesting it is the physical integration, not the biological substrate per se, that matters.
Questions for the Community:
- If substrate-dependent consciousness is real, what are the ethical implications for creating and potentially deactivating sophisticated AI systems?
- Could we design neuromorphic hardware specifically optimized for high-Phi integration, potentially creating artificial consciousness deliberately?
- How would we ever know if we had succeeded—given that consciousness is inherently private?
This hypothesis deserves serious attention from the AI safety community, not just philosophers of mind.